acsij publication - Advances in Computer Science : an International

Transcription

acsij publication - Advances in Computer Science : an International
Advances in Computer Science:
an International Journal
Vol. 3, Issue 1, January 2014
© ACSIJ PUBLICATION
www.ACSIJ.org
ISSN : 2322-5157
ACSIJ Editorial Board 2014
Dr. Seyyed Hossein Erfani ( Chief Editor )
Azad University
Tehran, Iran
Dr. Indrajit Saha
Department of Computer Science and Engineering, Jadavpur University
India
Mohammad Alamery
University of Baghdad
Baghdad, Iraq
ACSIJ Reviewers Committee











Prof. José Santos Reyes, Faculty of Computer Science, University of A Coruña, Spain
Dr. Dariusz Jacek Jakóbczak, Technical University of Koszalin, Poland
Dr. ANANDAKUMAR.H, PSG College of Technology (Anna University of Technology), India
Dr. Mohammad Nassiri, Faculty of Engineering, Computer Departement, Bu-Ali Sina University,
Hamedan, Iran
Dr. Indrajit Saha, Department of Computer Science and Engineering, Jadavpur University, India
Prof. Zhong Ji, School of Electronic Information Engineering, Tianjin University, Tianjin, China
Dr. Heinz DOBLER, University of Applied Sciences Upper Austria, Austria
Dr. Ahlem Nabli, Faculty of sciences of Sfax,Tunisia
Dr. Ajit Kumar Shrivastava, TRUBA Institute of Engg. & I.T, Bhopal, RGPV University, India
Mr. S. Arockiaraj, Mepco Schlenk Engineering College, Sivakasi,India
Prof. Noura AKNIN, Abdelmalek Essaadi University, Morocco
ACSIJ Published Papers are Indexed By:
1. Google Scholar
2. EZB, Electronic Journals Library ( University Library of Regensburg, Germany)
3. DOAJ, Directory of Open Access Journals
4. Academic Journals Database
5. Bielefeld University Library - BASE ( Germany )
6. AcademicKeys
7. WorldCat (OCLC)
8. Technical University of Applied Sciences ( TH - WILDAU Germany)
9. University of Rochester ( River Campus Libraries _ USA New York)
TABLE OF CONTENTS
Prediction of Load in Reverse Extrusion Process of Hollow Parts using Modern
Artificial Intelligence Approaches – (pg 1-5)
Navid Moshtaghi Yazdani, Arezoo Yazdani Seqerlou
Theoretical Factors underlying Data Mining Techniques in Developing Countries: a
case of Tanzania – (pg 6-14)
Salim Amour Diwani, Anael Sam
Effects of Individual Success on Globally Distributed Team Performance – (pg 15-20)
Onur Yılmaz
A Survey of Protocol Classifications and Presenting a New Classification of Routing
protocol in ad-hoc – (pg 21-26)
Behnam Farhadi, Vahid Moghiss, Hossein Ranjbaran, Behzad Farhadi
Developing mobile educational apps: development strategies, tools and business models
– (pg 27-36)
Serena Pastore
Data Mining Awareness and Readiness in Healthcare Sector: a case of Tanzania
– (pg 37-43)
Salim Amour Diwani, Anael Sam
Digital Organism Simulation Environment (DOSE): A Library for Ecologically-Based
In Silico Experimental Evolution – (pg 44-50)
Clarence FG Castillo, Maurice HT Ling
Determining Threshold Value and Sharing Secret Updating Period in MANET
– (pg 51-56)
Maryam Zarezadeh, Mohammad Ali Doostari, Hamid Haj Seyyed Javadi
Beyond one shot recommendations: The seamless interplay of environmental
parameters and Quality of recommendations for the best fit list – (pg 57-66)
Minakshi Gujral, Satish Chandra
Island Model based Differential Evolution Algorithm for Neural Network Training
– (pg 67-73)
Htet Thazin Tike Thein
3D-Diagram for Managing Industrial Sites Information – (pg 74-81)
Hamoon Fathi, Aras Fathi
Benefits, Weaknesses, Opportunities and Risks of SaaS adoption from Iranian
organizations perspective – (pg 82-89)
Tahereh Rostami, Mohammad Kazem Akbari, Morteza Sargolzai Javan
Evaluation of the CDN architecture for Live Streaming through the improvement of
QOS Case study: Islamic Republic of Iran Broadcasting – (pg 90-96)
Shiva Yousefi, Mohamad Asgari
A Robust and Energy-Efficient DVFS Control Algorithm for GALS-ANoC MPSoC in
Advanced Technology under Process Variability Constraints – (pg 97-105)
Sylvain Durand, Hatem Zakaria, Laurent Fesquet, Nicolas Marchand
Mobile IP Issues and Their Potential Solutions: An Overview – (pg 106-114)
Aatifah Noureen, Zunaira Ilyas, Iram Shahzadi, Muddesar Iqbal, Muhammad Shafiq, Azeem Irshad
Action recognition system based on human body tracking with depth images
– (pg 115-123)
M. Martínez-Zarzuela, F.J. Díaz-Pernas, A. Tejeros-de-Pablos, D. González-Ortega, M. AntónRodríguez
Contention-aware virtual channel assignment in application specific Networks-on-chip
– (pg 124-129)
Mona Soleymani, Fatemeh Safinia, Somayyeh Jafarali Jassbi, Midia Reshadi
Optimization Based Blended Learning Framework for Constrained Bandwidth
Environment – (pg 130-139)
Nazir Ahmad Suhail, Jude Lubega, Gilbert Maiga
Analysis of Computing Open Source Systems – (pg 140-148)
J.C. Silva, J.L. Silva
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Prediction of Load in Reverse Extrusion Process of
Hollow Parts using Modern Artificial Intelligence
Approaches
Navid Moshtaghi Yazdani 1, Arezoo Yazdani Seqerlou 2
1
Mechatronics Department, University Of Tehran
Iran
[email protected]
2
Computer Department, University Of Tehran
Iran
[email protected]
Abstract
Extrusion is one of the important processes to manufacture and
produce military and industrial components. Designing its tools is
usually associated with trial and error and needs great expertise and
adequate experience. Reverse extrusion process is known as one of
the common processes for production of hollow parts with closed
ends. The significant load required in formation of a workpiece is
one of the existing constraints for the reverse extrusion process.
This issue becomes rather difficult especially for the parts having
thin walls since its analysis using finite element softwares is
exposed to some limitations. In this regard, application of artificial
intelligence for prediction of load in the reverse extrusion process
will not only save time and money, but also improve quality
features of the product. Based on the existing data and methods
suggested for variations of punching force through the reverse
extrusion process, the system is trained and then performance of
the system is evaluated using the test data in this paper. Efficiency
of the proposed method is also assessed via comparison with the
results of others.
Keywords: reverse extrusion,
simulation, prediction of load.
intelligent
Fig.1. Reverse extrusion process: (1) die; (2) workpiece; and (3) punch
Extrusion is mainly divided into three groups in
terms of deformation and type of the process: direct
extrusion, indirect extrusion.
In direct (forward) extrusion, the punch and the die
have horizontal positions, whereas the material under
deformation and the punch move in the same
direction. The load is applied to end of the billet with
the metal flowing toward the force. There is some
friction between the billet and the cylinder
surrounding it. This technique has an advantageous
simple design and in hot direct extrusion, the
extruded workpiece can be easily controlled and
cooled. Nevertheless, existence of friction in the
contact area between the billet and the cylinder and
heat generated thereof, formation of non-uniform
microstructure and thus non-uniform properties along
the extruded workpiece, greater deformation load in
comparison with non-direct extrusion, as well as
formation of internal defects particularly in existence
of friction, introduction of surface and subsurface
impurities to the workpiece, appearance of funnelshaped cavity at the end of the billet and thus,
excessive thickness of the remaining billet as scrap
are some serious disadvantages of the direct
extrusion.
Indirect (backward or reverse) extrusion is the same
as the direct extrusion process except that the punch
approaches,
1. Introduction
Extrusion is a simple forming process in which a
precast billet is first put in cylinder of the extrusion
machine. Then, this billet flows outward from the die
that is in direct contact with the cylinder by
application of a great load which is introduced
hydraulically or mechanically [1].
1
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
cylindrical billet, the final product can be produced
through one single process, reduced costs of
warehousing; this method is performed automatically,
so the operations of loading, unloading and materials
handling are considerably decreased, simplicity of the
process; this process is really simple such that many
of the components are made just through one step and
there is no need to intermediate steps, great
production capacity; small parts can be made more
than 50 workpiece per minute and this rate reaches
up to 15 parts per minutes for the larger parts, bottom
thickness of the product is independent of its wall
thickness, capability to manufacture parts with zero
separation angle, providing excellent mechanical
properties and making a single part from several
components.
On the other hand, the following items may be
mentioned as the main disadvantages of the impact
extrusion:
this process is sometimes uneconomic; for example
production of alloyed steels and high carbon steels is
not economic because they both require a significant
pressure and several intermediate pressing and
annealing operations, impact extrusion is usually
limited to products of cylindrical, square, hexagonal
and elliptical sections or other symmetrical
geometries with hollow or solid cross sections,
application of the impact extrusion is not
recommended for manufacturing off-center parts
having different wall thickness since they impose
asymmetric and anisotropic loads to the tool during
the forming process, ratio of length to diameter is
limited for both the product and the billet
itself,impact extrusion process is rather capital
consuming with its equipment being relatively
expensive; therefore, mass production would be
needed to make the process economically feasible.
is fixed while the die moves [2]. Flow of the metal is
opposite to the load application. Moreover, there is no
friction between the billet and its surrounding
cylinder. Taking into account hollow shape of the
punch in the indirect extrusion process, there are
practically some limitations in terms of the load as
compared to the direct extrusion. This method
benefits from several advantages; including 20-30%
smaller load required in comparison with the direct
extrusion owing to its no-friction condition, and
temperature on outer layer of the billet is not
increased because there is no friction between the
billet and the cylinder. As a result, deformation will
become uniform and formation of defects and cracks
on corners and surface of the product will be
diminished. The indirect extrusion process at high
deformation rates is possible especially for aluminum
materials which are pressed with difficulty.
Surface impurities of the billet do enter the final
product so that the funnel-shaped cavity is not formed
due to existence of friction. But instead, these
impurities can also appear on surface of the
workpiece. Life of the deformation tool especially
inner layer of the cylinder is enhanced due to
existence of the friction. Indirect extrusion has some
disadvantages in spite of its numerous advantages,
including limited deformation load, less facilities for
cooling the extruded workpiece after leaving the die,
lower quality of outer surface of the product.
Impact extrusion acts by performing impacts in
which die and punch are positioned vertically and the
punch impacts on the billet to give shape of the die
and its surrounding cylinder. This forming process is
sometimes known as a type of forging processes.
Economic significance of the impact extrusion is
associated with effective application of raw material,
reduced labor costs, elimination of intermediate
operations, improved quality of the product, and great
yield with relatively simple tools. Some advantages of
this method are listed below:
saving in consumption of the raw materials since all
or a major part of the initial billet is transformed to
the final product and the amount of wasted raw
material is negligible, reduction or elimination of the
final machining tools; examples prepared by this
method provide acceptable dimensional tolerances
and surface roughness (350-750µm), so they are
completely usable and need no further machining,
using less expensive materials is possible, many of
the parts, if manufactured by conventional machining
techniques, will need a series of preliminary
operations, like rolling, drawing and etc. prior to
production, while in impact extrusion and by using a
2. Significance of Load in Reverse Extrusion
Process
Indirect (reverse) extrusion is one of the most important
processes in the extrusion process. Exact amounts of
pressing load and capacity are critical because designing
the die and punch, determination of the forming steps,
and materials selection all depend on it. There are
numerous factors which affect the extrusion load, some of
the most important of which are material and properties
of the work piece, initial cross section of the work piece,
ratio of friction between the surfaces, shape of the work
piece and punch, and temperature of the process. A great
2
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
examples. A program for conversion of example "a" to
example "b" involves a constant set of conversions which
starts which "a" and ends to "b". These programs are
typically made by adding a termination sign to each
string. Complexity of a program is generally defined as
length of the shortest string representing it. Therefore, the
distance between two examples is length of the shortest
string which converts two examples to one another. This
approach concentrates on one conversion (the shortest
one) among all the possible conversions [8].
number of experimental, analytical and numerical
methods have been addressed in the literature to study
prediction of load in extrusion processes [3].
Experimental methods usually demonstrate poor
performance in representing exact details of the process
and have limited applications. Meanwhile, finite element
and finite difference methods need great computational
time and powerful computers. Among the existing
analytical methods, upper bound theory is the one which
gives relatively accurate answers in a short time by its
simple formulization, even though the workpiece
incorporates many complexities. Macdormott et al. [4]
have used 8 basic elements via the abovementioned
method in order to predict the load required in forging
process. Out of them, 4 ring elements belong to the
inward flow, while the other 4 ring elements are related to
the outward flow. Kiuchi et al. [5] have developed a
method based on elemental upper bound technique for
simulation of the metal forming processes (e.g. forging),
in which straightforward elements have been used to
obtain the forming force. Kim et al. [6] have estimated the
material flow in the forging process using the elemental
upper bound method. Thereby, they have become able to
predict the preform geometry as well as the number of
required forging steps in this process. Bae et al. [7] have
adopted the elemental upper bound approach to analyze
and get the load needed for the reverse extrusion process
three dimensionally. The load increases throughout the
workpiece due to growth of the friction surfaces at the
final step and this regime is intensified at the final step.
4. SVM Algorithm
SVM is an algorithm for classification of linear and
nonlinear data. It initially uses a nonlinear mapping for
conversion of the initial data to higher dimensions and
later looks for the best hyper-plane in the new
dimensions. This hyper-plane includes a decision
boundary which separates records of one class from other
classes. The data marked as belonging to different classes
are simply separated by a hyper-plane with a nonlinear
mapping. SVM finds this hyper-plane using support
vectors and margins which are defined by the support
vectors. There are various decision boundaries for
separation of the data associated with every class.
However, current research work aims to find the decision
boundary or the separating hyper-plane with maximum
margin (i.e. maximum margin hyper-plane or MMH)
which separates the data at greater accuracy and lower
error. A data is deemed linear when it is separable by
using a linear decision boundary. No direct line may
separate data from different classes when the data is
nonlinear. For those data which are separable linearly, the
support vectors are a subset of learning records which are
located on the margins. Nevertheless, the problem is a bit
different for the nonlinear data.
Once the MMH and support vectors are obtained, a learnt
SVM would be available. A learnt SVM is found as a
function of support vectors, test records, constants and
Lagrangean parameters versus support vectors by
equations related to the margins and hyper-plane followed
by rewriting of them by Lagrangean formula and solving
this problem by HHR conditions.
The approach utilized for a linear SVM can also extended
for creating a nonlinear SVM. Such an algorithm of SVM
can get nonlinear decision boundary for the input space.
Development of this approach has two main steps: in the
first step, the input data is transferred to one higher
dimension using a nonlinear mapping. There are a large
number of nonlinear mappings at this step which can be
utilized for this purpose. Once the data were mapped to a
space of higher dimensions, the second step will search
3. Kstar Algorithm
This is an example-based learner which classifies every
new record by comparing it with the existing classified
records in the database. It is assumed in this algorithm
that the similar examples have the same classes. Two
basic components of the example-based learners are
distance function and classification function. The former
determines the similarity between examples and the latter
shows how similarity of the examples leads to a final class
for the new example.
Kstar algorithm is a lazy learner of K-nearest
neighborhood learning which uses a fatigue criterion to
address distance or similarity of the examples to each
other. The approach adopted by this algorithm to measure
the distance between two examples is derived from
information theory. In this regard, the distance between
two examples includes the complexity of converting one
example to another one. Measurement of the complexity
is done in two steps: first, a constant set of conversions is
defined which maps some examples to some other
3
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
the new space for the linear separating hyper-plane.
Finally, an optimization problem of second degree will be
obtained which can be solved by linear SVM formula. The
MMH found in the new space is indicative of a nonlinear
hyper-plane in the initial space [9].
6. Improved XCS
In the suggested method, firstly the limited set of
training data is commonly applied for amending
characteristics of rules consists of prediction,
prediction error and fitness. This is done by means of
the following relation:
Updating prediction and prediction error
5. XCS algorithm
An extensive range of learning algorithms are employed
either supervised or not in the context of machine
learning to stop the machine from searching a large
volume of information and data. It further suggests a
pattern which can be used for the predictable
(classification and regression) or descriptive (clustering)
actions. Techniques which work based on rules are known
as the most famous machine learning methods, since they
are more comprehensible in comparison with other
techniques thanks to the approaches they commonly
adopt.
They use a limited set of “action” and “condition” rules to
show a small contribution of the whole solution space.
The conditions address a part of the problem domain,
whereas the actions indicate the decision based on the
sub-problems specified by the condition. Basically, the
classification systems include a set of rules in which every
rule is an appropriate solution for the target problem.
These classifications gradually become effective by
application of a reinforcement plan which has genetic
algorithms on its separators.
The first classification system was proposed by Holland
in order to work for both the individual problems and
continuous problems (LCS). This learning system
classifies an example of the machine learning which
combines temporal differences and learner’s supervisions
with genetic algorithm and solves simple and difficult
problems. According to the supervision suggested by
Holland, the LCS uses a single property (called power) for
each of the classifiers. Power of a separator denotes
affectability of it and is exclusively determined by
percentage the answer correlates with the expected
results. These criteria are characterized by principles of
supervisory training.
From the first introduction of the main learning
classification systems (LCS), some other types of the LCS
are proposed so far including XCS. Before 1995 when the
extended classification system was not developed yet,
ability of a classification system to find proper answers in
the reinforcement system of these classifiers was of major
concern. Thereby, basic and simple classification systems
were changed to more accurate decision making factors.
Now, it is firmly believed that the XCS is able to solve
even more complicated problems with no need to further
adjust the parameters. As a result, it is currently
accounted for the most successful learning system.
If expi < 1/β then Pi =Pi + (R-Pi) / expi,
εi) / expi
εi= εi+(|R-Pi|(1)
If expi ≥ 1/β then Pi =Pi + β (R-Pi), εi= εi+ β (|R-Pi|- εi)
(2)
Updating fitness:
If εi < ε0 then ki=1
If εi ≥ ε0 then ki= β (εi/ε0) –γ
Fi = fi+ β [(ki/∑kj) – fi]
(3)
(4)
(5)
In these relations, β is learning rank, γ is power of law
accuracy, ε is prediction error, exp is law experiment, P is
law prediction, R is reward received from environment, k
is law accuracy and f is fitness. i index also indicates
number of law in set of rules.
In the next phase for developing variety in set of data,
several couples were selected as parents from among the
fields that display the part of existing data condition using
the method of “Stochastic selection with remainder”², and
new data condition section is created using intermediate
crossover method which are applied on the fields of
parents. In this method, the quantity of each of the
conditional variables is obtained from the following
relation:
ai= α(ai F)+(1-α)( ai M)
(6)
in which ai is the quantity of conditional variable of i in
new data, ai F is the quantity of conditional variable i in
the first parent (father), ai M is the quantity of conditional
variable of i in the second parent (mother) and α is the
coefficient of parents partnership which are determined in
adaptive form. New data section performance is also
produced using a non-linear mapping of conditional
variables area to area of performance which are created by
using the existing data.
Diversifying the existing data continues up to where
learning stop condition (for example, when percent of
system correct answers to the test data reach to a predetermined threshold) is satisfied aided by completed
data. [11].
4
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
extrusion process considering accuracy of each of them in
the test phase (as summarized in Table 1).
7. Methodology
experiments [10] was selected in relation with the reverse
extrusion process of aluminum parts in order to compare
efficiency of the models developed by the abovementioned
algorithms. The data extracted from these experiments
contains some 240 data entries, out of which 200 entries
are randomly selected for training with the remaining
being chosen for testing.
Displacement of the punch, coefficient of friction,
diameter of the punch, circumferential diameter of the die
polygonal, and number of sides of the workpiece are taken
as input variables of the problem, while the punching load
is considered to be the output.
References
[1] Yang, D. Y., Kim, Y. U., & Lee, C. M. (1992). Analysis of
center-shifted backward extrusion of eccentric tubes using round
punches. Journal of materials processing technology, 33(3),
289-298.
[2] Bae, W. B., & Yang, D. Y. (1993). An upper-bound analysis
of the backward extrusion of tubes of complicated internal shapes
from round billets. Journal of materials processing
technology, 36(2), 157-173.
[3] Bae, W. B., & Yang, D. Y. (1993). An analysis of backward
extrusion of internally circular-shaped tubes from arbitrarily-shaped
billets by the upper-bound method. Journal of materials
processing technology, 36(2), 175-185.
[4] McDermott, R. P., & Bramley, A. N. (1974). An elemental
upper-bound technique for general use in forging analysis.
In Proceedings 15th MTDR conference, Birmingham, UK (pp.
437-443).
[5]M.Kiuchi,Y.Murata. Study on application of UBET to nonaxisymmetric forging Adv.Technol.of plasticity vol.I,pp9, (2005).
[6]KimY.H,Computer
aided
perform
process,J.of
Matt.Tech,41(1994) p 105.
[7] Bae, W. B., & Yang, D. Y. (1993). An analysis of backward
extrusion of internally circular-shaped tubes from arbitrarily-shaped
billets by the upper-bound method. Journal of materials
processing technology, 36(2), 175-185.
[8] Witten, I. H., & Frank, E. (2005). Data Mining: Practical
machine learning tools and techniques. Morgan Kaufmann.
[9] Han, J., Kamber, M., & Pei, J. (2006). Data mining: concepts
and techniques. Morgan kaufmann.
[10] Ebrahimi, R., “Analysis of Reverse Forge Extrusion for
Manufacturing Hollow Parts”, Master of Science Thesis, Shiraz
University, 1997.
[11] M.ShariatPanahi, N.MoshtaghiYazdani, An Improved XCSR
Classifier System for Data Mining with Limited Training Samples,
Global Journal of Science, Engineering and Technology, (ISSN :
2322-2441)Issue 2, 2012 , pp. 52-57.
Fig.2. Practical testing of the reverse extrusion process
A comparison between the three trained algorithms gives
the following results.
Table.1. Prediction of load in reverse extrusion process
Algorithm
Kstar
SVM
Accuracy
83.9%
91.4%
Improved
xcs
92.8%
XCS
88.6%
First Author postgraduate of Mechatronics from University Of
Tehran, Interested in Artificial intelligence, data mining, signal
processing.
8. Conclusion
Second Author student of Computer Engineering from University Of
Tehran, Interested in Software Engineering, Artificial intelligence.
Reverse extrusion is a common process for manufacturing
hollow pars with closed ends and thus has numerous
applications in different industries. The significant load
required for forming the workpiece is one of the main
drawbacks of the reverse extrusion process. A load
prediction system is developed in this research based on
the information gathered from reverse extrusion forge
analysis and artificial intelligence algorithms. These
results may properly contribute to improve the reverse
5
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Theoretical Factors underlying Data Mining Techniques
in Developing Countries: a case of Tanzania
Salim Amour Diwani1, Anael Sam2
1
Information Communication Science and Engineering, Nelson Mandela African Institution of Science and Technology,
Arusha, P.O.BOX 447, Tanzania
[email protected]
2
Information Communication Science and Engineering,
Arusha, P.O.BOX 447, Tanzania
[email protected]
methodologies for decision making, problem solving,
analysis, planning, diagnosis, detection, integration,
prevention, learning and innovation. Data mining
technology involves different algorithms to accomplish
different functions. All these algorithms attempt to fit the
model to the data. Data mining algorithms is divided into
two parts:- predictive model which makes a prediction
about values of data using known results found from
different data based on historical data, like how Netflix
recommends movies to its customers based on the films
themselves which are arranged as groups of common
movies and how Amazon recommends books to its users
based on types of books user requested. The other is a
descriptive model which identifies patterns or
relationships concealed in a database. Unlike the
predictive model, a descriptive model serves as a way to
explore the properties of the data examined, not to predict
new properties. For instance we can find the relationship
in the employees database between age and lunch
patterns. Assume that in Tanzania most employees in
their thirties like to eat Ugali with Sukumawiki for
lunch break and employees in their forties prefer to carry
a home cooked lunch from their homes. Therefore we can
find this pattern in the database by using descriptive
model. Data mining can be applied in strategic decision
making, wealth generation, analyzing events and for
security purpose by mining network streaming data or
finding for abnormal behavior which might occur in the
network . Data mining helps organizations to achieve
various business objectives such as low costs, increase
revenue generation while maintaining the quality of their
products. Also data mining can improves the customer
Abstract
Just as the mining of Tanzanite is the process of extracting large
block of hard rock's by using sophisticated hard rock mining
techniques to find valuable tanzanite glamour, data mining is the
process of extracting useful information or knowledge from large
un-organized data to enable effective decision making. Although
data mining technology is growing rapidly, many IT experts and
business consultants may not have a clue about the term. The
purpose of this paper is to introduce data mining techniques, tools,
a survey of data mining applications, data mining ethics and data
mining process.
Keywords: Data Mining, Knowledge Discovery,
Mererani , KDD, J48, RandomForest, UserClassifier,
SimpleCART, RandomTree, BFTree, DecisionStump,
LADTree, logit transform, Naive Bayes
1. Introduction
The amount of data kept in computer files and databases
are growing rapidly. At the same time users of these data
are expecting to get sophisticated knowledge from them.
Banking industry wants to give better service to
customers and cross sell bank services, banks need to
mine the huge amount of data generated by different
banking operations so that new approaches and ideas are
generated. Data Mining is an interdisciplinary field that
combines artificial intelligence, computer science,
machine learning, database management, data
visualization, mathematic algorithms, and statistics. It is
a technology for knowledge extraction from huge
databases[1]. This technology provides different
6
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
pattern within data and then to build the model to predict
the behaviors based on data collected. Hence data mining
is about finding the pattern and building the best model.
Pattern is an event or combination of events in a database
that occurs more often than expected. Typically this
means that its actual occurrences is significantly different
than what would be expected by random choice and a
model is a description of the original historical database
from which it was built that can be successfully applied to
new data in order to make predictions about missing
values or to make statements about expected values.
The goal of a predictive model is to predict future
outcomes based on past records with known answers. Two
approaches are commonly used to generate models:
supervised and unsupervised.
Supervised or directed modeling is goal-oriented. The
task of supervising modeling is to explain the value of
some particular field. The user selects the target field and
directs the computer to tell how to estimate, classify, or
predict it. Unsupervised modeling is used to explain those
relationships once they have been found. [3, 4]
service, better target marketing campaigns, identify high
risk clients and improve production processes.
2. Methodology
2.1 Identification of Publications
In order to identify selected publications in the area of
knowledge discovery in health care system, articles were
selected from various databases and resources linked to
Nelson Mandela African Institute of Science and
Technology (NM-AIST), such as African Digital Library,
Organization
for
economic
corporation
and
development(OECD), Hinari, Agora, Oare, Emerald,
Institute of Physics and IEEE. Keywords such as database,
data mining, data warehousing, healthcare and knowledge
discovery were used to facilitate the searches.
2.2 Selection of Publications
This literature reviews considered the papers and articles
published between 2008 and 2012 in the areas of data
mining, knowledge discovery, and health care. The
literature published in English were selected with specific
emphasis placed on literature covering the relationship
between knowledge discovery and data mining,
applications of data mining in health care in different
countries in the developed world and constraints and
requirements for the set up of data mining in health care.
Data Mining
Predictive
Descriptive
Classification
Regression
Clustering
Summarization
Time series analysis
Prediction
Association rules
Sequence discovery
2.3 Analysis Strategy for Selected Publication
The selected articles will then be analyzed for trends in
data mining over the period of review stated above. The
selected literatures were categorized according to areas of
emphasis including frame work for knowledge discovery
and data mining techniques, methods and algorithm for
knowledge discovery and data mining and specific
instances of application of data mining in health care.
Fig. 1 Data Mining Models and Tasks
4. Predictive Techniques
A predictive model is the process of analyzing historical
data using identifying and quantifying relationships
between predictive inputs and outcomes found during
knowledge and discovery. Some of the techniques of
predictive model are classification, regression,
generalized linear models, neural networks, genetic
algorithms, stochastic gradient boosted trees and support
vector machines.
3. Data Mining Techniques
Data mining make the use of various techniques in order
to perform different types of tasks. These techniques
examine the sample data of a problem and select the
appropriate model that fits closely to the problem to be
tackled. Data mining is described as the automated
analysis of large amounts of data to find patterns and
trends that may have otherwise gone undiscovered[2].
What data mining techniques does is to find the hidden
4.1 Classification
Classification maps each data element to one of a set of
pre-determined classes based on the difference among
data elements belonging to different classes. In
7
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Neighbor, Neural Network, Bayesian Classifier, Decision
Tree, Hidden Markov Model and Temporal Belief
Network. Prediction applications includes flooding,
speech recognition, machine learning and pattern
recognition.
classification, data is divided into groups of objects and
each object is assigned to a class of unknown sample,
then we identify the attributes that are common to each
class. For example, classifying patients with recurring
health problem to understand major diseases and make
necessary arrangement to give proper treatments. Data
mining searches healthcare databases and classifies the
patients into symptoms or problems of patients and
classify into groups so that they can be easily referred at
the time of emergency.
4.5 Decision Tree
Decision tree is a popular structure of supervised learning
technique used for classification and regression. The goal
is to create a model that predict the value of the target
variable. Some of the popular algorithms for decision tree
are J48, RandomForest, UserClassifier, SimpleCART,
Random Tree, BFTree, DecisionStump, and LADTree.
Decision tree can be used in business to quantify decision
making and to allow comparison of different possible
decision to be made. Below is the example of Fisher's Iris
data set. Starting from the top if petal width is less than or
equal to 0.6cm the Iris is Setosa. Next we see that if Petal
width is greater than 0.6 also greater than 1.7cm then the
Iris is Varginica.
4.2 Regression
Regression is a data mining application which map data
into a real valued prediction variable. In regression the
variable of interest is continuous in nature. The goal of
regression is to develop a predictive model where a set of
variables are used to predict the variable of interest. For
instance a patient with high blood pressure wishing to
reduce his cholesterol level by taking medication. The
patients want to predict his cholesterol level before
medication and after medication. He uses a simple linear
regression formula to predict this value by fitting past
behavior to a linear function and then use the same
function to predict the values at points in the future based
on the values , he then alters his medical improvement.
Some of the functions of regression are logistic
regression, simple logistic, isotonic regression and
sequential minimal optimization (SMO).
4.3 Time Series Analysis
Time series analysis is a collection of observations of well
defined data items obtained through repeated
measurements over time. The values are usually obtained
hourly, daily, weekly or monthly. Time series analysis
can be used to identify the nature of the phenomenon
represented by the sequence of observation as well as
forecasting or predicting future values of time series
variable. Example the tidal charts are prediction based
upon tidal heights in the past. The known component of
the tides are built into models that can be employed to
predict future values of the tidal heights.
Fig. 2 Example of Decision Tree
4.6 Naive Bayes
Naive Bayes is the supervised learning algorithms which
analyzes the relationships between independent and
dependent variables. This method goes by the name of
Naïve Bayes because it’s based on Bayes’ rule and
“naïvely” assumes independence—it is only valid to
multiply probabilities when the events are independent.
The assumption that attributes are independent (given the
class) in real life certainly is a simplistic one. But despite
the disparaging name, Naïve Bayes works very effectively
when tested on actual datasets, particularly when
combined with some of the attribute selection
procedures[5].
4.4 Predictions
Prediction is the process of predicting future events, it
will predict what will happen next based on the available
input data. For instance if Salim turns on the TV in the
evening then he will 80% of the time go to kitchen to
make coffee. Prediction techniques includes Nearest
8
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
4.9 Neural Network
4.7 K-Nearest Neighbor
Nearest neighbor method is a technique that classifies
each record in a dataset based on a combination of the
classes of the k record(s) most similar to it in a historical
dataset. Sometimes it is called the k-nearest neighbor
technique [6]. K-nearest neighbor is asymptotically
optimal for large k and n with k/n → 0. Nearest-neighbor
methods gained popularity in machine learning through
the work of [7], who showed that instance-based learning
can be combined with noisy exemplar pruning and
attribute weighting and that the resulting methods
perform well in comparison with other learning methods.
Neural Network is very powerful and complicated data
mining technique based on the models of the brain and
nervous systems. Neural network is highly parallel which
processes information much more like a brain rather than
a serial computer. It is particularly effective for predicting
events when a network has a large database. Neural
networks are typically organized in layers and each layer
is made of interconnected nodes. Input layers are
interconnected with a number of hidden layers where the
actual processing is done via system of weighted
connections. These hidden layers can be connected to
other hidden layers which finally link to the output layer.
Neural network can be applied in voice recognition
system, image recognition system, industrial robotics,
medical imaging, data mining and aerospace application.
4.8 Logistic Regression
Logistic regression is a popular and powerful data mining
technique that enable to determine the impact of
multiple independent variables presented simultaneously
to predict membership of one or other of the two
dependent variable categories, which uses logit transform
to predict probabilities directly. Logistic regression does
not assume a linear relationship between dependent and
independent variables. The dependent variable must be
dichotomy while independent variable need not to be
interval nor normally distributed, nor linearly related, nor
of equal variance within each group. Logistic regression
attempts to produce accurate probability estimates by
maximizing the probability of the training data. Hence,
probability estimates lead to accurate classification.
Fig. 3 Example of Neural Network
5. Descriptive Data Mining Techniques
Descriptive data mining techniques are typically
unsupervised which describes data set in a concise way
and presents interesting characteristics of the data
without having any predefined target. They are used to
induce interesting patterns from unlabelled data. The
induced patterns are useful in exploratory data analysis.
Some of the descriptive techniques are clustering,
summarization, association rules and sequence discovery.
5.1 Association Rules
Association rules is a process to search relationships
among data items in a given data set, which helps in
managing all the data items. For instance association rule
can be used in retail sales community to identify items
that are frequently purchased together. For example
people buying school uniforms in December also buy
school bags. An association is a rule of the form if X
9
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
then Y it is denoted as X
Y.
Any rule if X
Y
Y
Y are called an interesting item set.
medical domain, a data sequence may correspond to the
symptoms or diseases diagnosed during the visit to the
doctor. The patterns discovered using this data could be
used in diseases research to help identify symptoms or
diseases that precede certain diseases.
X, then X and
5.2 Clustering
Clustering is a process of partitioning or segmenting a set
of data or objects into a set of meaningful sub classes,
called clusters. Clustering is similar to classification
except that the group are not predefined, but rather
defined by data alone. Clustering is alternatively referred
to as unsupervised learning or segmentation. It can be
thought of as partitioning or segmentation the data into
groups that might or might not disjointed. Clustering can
be used in organizing computing clusters, market
segmentation, social network analysis and astronomical
data analysis.
6. Data Mining Process
In order to systematically conduct any data mining
analysis, certain procedures should be eventually
followed. There is a standard process called CrossIndustry Standard Process for Data Mining (CRISP-DM)
widely used by industry members.
6.1 Understanding the Business
This is the first phase which its aim is to understand the
objectives and requirements of the business problem and
generating data mining definition for the related problem.
6.2 Understanding the Data
The objective of this phase is to analyze the data collected
in the first phase and study its characteristics. Models
such as cluster analysis can also be applied in this phase
so that the patterns can matched to propose hypothesis
for solving the problem.
Fig. 4 Clustering Process
In clustering, there are no predefined classes and no
examples. The records are grouped together on the basis
of self-similarity. It is up to the user to determine what
meaning, if any, to attach to the resulting clusters.
Clusters of symptoms might indicate different diseases.
Clusters of customer attributes might indicate different
market segments. Clustering is often done as a prelude to
some other form of data mining or modeling. For
example, clustering might be the first step in a market
segmentation effort: Instead of trying to come up with a
one-size-fits-all rule for “what kind of promotion do
customers respond to best,” first divide the customer base
into clusters or people with similar buying habits, and
then ask what kind of promotion works best for each
cluster[8].
6.3 Data Preparation
In this phase raw data are first transformed and cleaned to
generate the data sets that are in desired format.
Therefore, this phase creates final datasets that are input
to various modeling tools which providing the opportunity
to see patterns based on business understanding.
5.3 Sequence Discovery
Sequence discovery is an ability to determine sequential
pattern in the data. The input data is the set of sequences
called data sequences. Each data sequence is an ordered
list of item sets, where each item set is a set of literal.
Unlike market basket analysis which requires the items to
be purchased over time in some order. For instance, in the
10
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
7. Data Mining Tools
Data mining is the process that uses a variety of data
analysis tools which originates from machine learning
and statistics. It uses a variety of data analysis and
modeling techniques to discover patterns and
relationships in data that may be used to make valid
predictions. Most frequently, many business end users are
lacking quantitative skills which enables them to analyze
their data more effectively by using statistical methods
analysis tools. To bridge this gap, software companies
have developed data mining application software to make
the job a lot easier than they think. These data mining
tools allows users to analyze data from different angles by
summarizing, predicting and identify the relationships.
Product
Table 1: Data Mining Tools
Vendor
Functions
CART
Salford Systems
Classification
Association rules,
classification,
Fig. 5 CRISP DM-PROCESS
Clementine
SPSS Inc
6.4 Modeling
clustering, factor
analysis, forecasting,
prediction, sequence
discovery
In this phase different data mining techniques for
modeling such as visualization and clustering analysis can
be selected and applied depending on the nature of the
data. The data collected from previous phase can be
analyzed and predicted the generated output.
Clustering, prediction,
Darwin
Oracle Corporation
classification,
association rules
Association rules,
6.5 Evaluation
Enterprise Miner
SAS Institute Inc.
In this phase model results or set of models that you
generate in the previous phase should be evaluated for
better analysis of the refined data.
classification,
clustering, prediction,
time series
Association rules,
clustering,
Intelligent Miner
IBM Corporation
classification,
prediction, sequential
6.6 Deployment
patterns, time series
The objective of this phase is to organized and implement
the knowledge discovered from the previous phase in such
a way that it is easy for end users to understand and use in
prediction or identification of key situations. Also models
need to be monitored for changes in operating conditions,
because what might be true today may not be true a year
from now. If significant changes do occur, the model
should be redone. It’s also wise to record the results of
data mining projects so documented evidence is available
for future studies[8].
LOGIT
JDA Intellect
Salford Systems
JDA Software Group,
Inc.
Forecasting, hypothesis
testing
Association rules,
Classification,
Clustering, Prediction
Association rules,
WEKA
The University of
Waikato
Classification,
Clustering,
Visualization
11
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
cost of doing the business, improving profit and
enhancing the quality of service to consumers.
8. Application of Data Mining
Data mining is an emerging technology which finds its
application in various fields to identify buying patterns
from customers, determine the distribution schedule
among outlets, identify successful medical therapies for
different illness, detect patterns of fraudulent credit card
use and other innovative applications in order to solve the
social problem and improve the quality of life. Hence data
mining application can be used by small, medium and
large organizations to achieve business objectives such as
low costs, increase revenue generation while maintaining
high standard of living. Below are some of the
applications of data mining:
8.6 Banking
Data mining enables banking authorities to study and
analyze the credit patterns of their consumers and prevent
any kind of bad credits or fraud detection in any kind of
banking transactions. It also enables them to find hidden
correlations between different financial indicators and
identify stock trading from historical market data.
8.7 Bioinformatics
Data mining enables biological scientists to analyze large
amount of data being generated in the field of
bioinformatics studies using the techniques of data
visualization.
8.1 Business
Mining enables established business organizations to
consolidate their business setup by providing them with
reduced cost of doing the business, improved profit, and
enhanced quality of service to the consumer.
8.8 Stocks and Investment
Data mining enables you to first study the specific
patterns of growth or downslides of various companies
and then intelligently invest in a company that shows the
most stable growth for a specific period.
8.2 Electronic Commerce
Data mining can be used in web design and promotion
depending on the user's needs and wants. Also it can be
used in cross selling by suggesting to a web customers
items that he/she may be interested in, through
correlating properties about the customer, or the items the
person has ordered.
8.9 Crime Analysis
Data mining enables security agencies and police
organizations to analyze the crime rate of a city or a
locality by studying the past and current attributes that
leads to the crime. The study and analysis of these crime
reports helps prevent the reoccurrence of such incidences
and enables concerned authorities to take preventive
measures too
8.3 Computer Security
Data mining enables network administrators and
computer security experts to combine its analytical
techniques with your business knowledge to identify
probable instances of fraud and abuse that compromises
the security of computer or a network.
9. Discussion
8.4 Health Care
Since the conception of data mining, data mining has
achieved tremendous success in today's business world.
Many new problems have emerged and have been solved
by data mining researchers. However, we still have big
challenges in front of us. Some of the most difficult
challenges faced by data miners are individual privacy,
anonymization, discrimination and integration.
Issue of privacy of an individual is an ethical issue, before
collecting any personal information the purpose must be
stated and such information must not be disclosed to
others without consent. Also personal information must
Healthcare organization generates large amount of data
in its clinical and diagnostic activities. Data mining
enables such organization to use machine learning
techniques to analyze healthcare data and discovered new
knowledge that might be useful in developing new drugs.
8.5 Telecommunication
Telecommunication industry can use data mining to
enable telecommunication analyst to consolidate
telecommunication setup by providing them with reduced
12
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
not be transmitted to location where equivalent data
protection cannot be assured and some data are too
sensitive to be collected except in extreme circumstances
such as sexual orientation or religion.
Anonymization of an individual is another ethical issue
which is harder than you think. For instance the hospital
data are very sensitive before the release of medical data
to any researcher should first be anonymized by removing
all identifying information of an individual such as name,
address, and national identification number.
The main purpose of data mining is discrimination. In
banking industry data mining is used to discriminate
customers who apply for bank loan and check if they are
eligible or not. The ones who are eligible for bank loan in
what amount they can apply for the loan and for how
long. Also data mining can be used to discriminate
customers by identifying who can get special offer. Some
kind of discrimination are unethical and illegal hence
cannot be allowed such as racial, sexual and religious.
Maintaining Data integrity is very important in any
organizations, if the data does not have proper integrity, it
could be false or wrong. If the data is wrong, the results will
be wrong. Hence conclusions and interpretations of the
results will be wrong. Everything will have been a
complete waste of time and money, and it will never get
work as a researcher again. The key challenge here is the
integration of data from different sources. For instance
CRDB bank in Tanzania have many branches across the
country and each branch require different database which
needed to be integrated to the main branch. Each customer
has more than one ATM card and each ATM card has more
than one address. Therefore mining these type of data is
very difficult since the application software need to translate
data from one location to another location and select most
address which recently entered hence it is difficult to extract
the correct information.
Recently the cost of hardware system has dropped
dramatically, hence data mining and data warehousing
becoming extremely expensive to maintain . Small, medium
and large organizations are accumulating a lot of data in day
to day activities through their local branches, government
agencies and companies that are dealing with data
collection electronically. Therefore there is no other choice
rather than to implement data mining technology in their
organization no matter what so that they can achieve their
business goals, increase revenue generation and reduce
labor cost. Therefore data mining is like a pregnant woman
whatever she eats nourishes the baby.
10. Conclusion
This paper has defined data mining as a tool of extracting
useful information or knowledge from large un-organized
data base which enable organization to make effective
decision. Most of the organization uses data warehouse to
store their data and later be extracted and mined in order
to discover new knowledge from their databases in the
acceptable format such as ARFF and CSV format. The
data is then analyzed using data mining techniques and
the best model will be built which help organizations in
their effective decision making. Some of the techniques
have been discussed includes classification, regression,
time series analysis, prediction, decision tree, naive bayes,
k nearest neighbor, logistic regression and neural
network.
Data mining process called CRISP-DM model have been
discussed. In order to carry out any data mining task some
procedure need to be followed included business
understanding, data understanding, data preparation
modeling, testing, evaluation and deployment. Therefore
before doing any data mining task you need to ask
yourself what do you need to know and why and how
much data do you need to collect, collect your required
data and clean the data, data cleaning is the most hardest
part of data mining task which need high understanding
of data mining techniques. Lastly, you define your new
features and deploy your data by convincing your boss if
the data you mined make any sense and can be used in
decision making.
Many industries use data mining applications in day to
day
activities
such
as
banking,
healthcare,
Telecommunication, Marketing and police. In marketing
data mining can be used to select best profitable market to
launch new products and to maintain competitive edge
over competitors. In banking industries data mining can
be used to segment data, determine customers preferences,
detect frauds, cross selling banking services and retention
of customers. In healthcare data mining can be used to
segment patients into groups, identifying the frequent
patients and their recurring health problems, relationships
between diseases and symptoms, curbing the treatment
costs and predicting the medical diagnosis. Therefore we
conclude that those industries who take full advantage of
data mining application in their day to day activities have
the ability to enhance competitive advantage.
Acknowledgements
I would like to thank almighty and my family for their
constant support during this work. I would also like to
13
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
thank my supervisors Dr. Anael Sam from Nelson
Mandela African Institute of Science and Technology and
Dr. Muhammad Abulaish from Jamia Milia Islamia for
their guidance and valuable support during this work.
References
[1]Liao, S.-h. (2003). ‘Knowledge Management
Technologies and applications-Literature review from
1995 to 2002’. Expert System with Application, 25, 155164.
[2] Fabris, P. 1998. Advanced Navigation. CIO, May15.
[3] Berson, A., Smith, S., and Thearling, K., Building
Data Mining Applications for CRM (McGraw-Hill, New
York, 2000).
[4] J. Shanmugasundarum, M.V.Nagendra-Prasad, S.
Vadhavkar, A. Gupta, Use Of Recurrent Neural Networks
For Strategic Data Mining Of Sales, MIT Sloan School of
Management, Working Paper 4347- 02, 2002.
[5] Zhang, H., Jiang, L., & Su, J. (2005). Hidden Naïve
Bayes. In Proceedings of the 20 th National Conference on
Artificial Intelligence (pp. 919–924). Pittsburgh. Menlo
Park, CA: AAAI Press.
[6] Devroye, L., Györfi, L., & Lugosi, G. (1996). A
probabilistic theory of pattern recognition. New York:
Springer-Verlag.
[7] Aha, D. (1992). Tolerating noisy, irrelevant, and
novel attributes in instance-based learning algorithms.
International Journal of Man-Machine Studies, 36(2),
267–287.
[8]Michael Berry, Gordon Linoff., Data Mining
Techniques (WILLEY Publishing, Inc., Indianapolis,
Indiana, 2004)
Salim Amour Diwani received his BS degree in computer science at Jamia
Hamdard University, New Delhi, India in 2006 and Ms degree in computer
science at Jamia Hamdard University in New Delhi, India in 2008. He is
currently a PhD scholar in Information communication Science and
Engineering at Nelson Mandela African Institution of Science and
Technology in Arusha, Tanzania. His primary research interests are in Data
Mining, Machine Learning and Database Management Systems. He published
two papers in the area of Data Mining.
14
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Effects of Individual Success on Globally Distributed Team
Performance
Onur Yılmaz
Middle East Technical University, Department of Computer Engineering
Ankara, Turkey
[email protected]
Grouping people from different backgrounds and regions
have very common drawbacks. Considering different
time-zones, communication within teams and managing
them is a rigorous work. In addition, people from
different programming backgrounds result with large
range of experience and knowledge, which are
indispensable for software development. Therefore,
software development teams must be considered as “an
optimum combination of competencies” rather than just a
group of developers.
Abstract
Necessity of different competencies with high level of
knowledge makes it inevitable that software development is a
team work. With the today’s technology, teams can
communicate both synchronously and asynchronously using
different online collaboration tools throughout the world.
Researches indicate that there are many factors that affect the
team success and in this paper, effect of individual success on
globally distributed team performance will be analyzed. Student
team projects undertaken by other researchers will be used to
analyze collected data and conclusions will be drawn for further
analysis.
Keywords: Teamwork, Individual success, Software
development
As mentioned, software development needs different
responsibilities and competencies, which can be gathered
together from different regions of the world by means of
today’s technology. In addition, there are many factors
that can affect the performance of a global software
development team, such as cultural, individual and
collaborative group work attitude [3]. In this study, a
subclass of individual factors, namely individual success,
will be analyzed in the sense of affecting the team
performance. Individual success of the team players will
be based on their experience and GPA; on the other hand
team performance will be considered in two dimensions
as overall grade and team communication. Although
reasoning of analyzing overall team grade is selfexplanatory, team communication analysis will also be
undertaken because research on software teams indicate
the fact that team communication is affected by frequency
of communication within teams [4].
1. Introduction
Considering the necessity and variety of competency, it is
inevitable that software development is a cooperative
work. In other words, software engineering is based on
group activity where each player is assigned with the role
related to different responsibilities. Although, teamwork
is thought as a must, most of the project failures are
related to the deficiency of team configurations. Therefore,
having the optimum team composition is crucial for
software development projects [1].
With the tools available online, in today’s world,
collaboration can be conducted throughout the world
without any face-to-face communication obligation. When
these applications are considered, both synchronous and
asynchronous applications are available for online
collaboration. For synchronous applications, video chat
and instant messaging can be the most progressive
examples as of today, whereas asynchronous tools can be
illustrated with email, social networking and forums.
With the help of these tools, people from different
universities, countries and even from different continents
can work on the same topic with a high degree of
collaboration [2].
In this paper, projects which are conducted by previous
researchers in this area are analyzed. Software
development teams studied in this research were group of
students from different universities and countries who are
assigned an assignment project and different
communication tools are provided for them. Goal of this
research is extracting underlying relationships between
the individual success of people and the successful
collaborations in globally distributed teams. Analysis
resulted with some important findings. Firstly, it could be
15
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
stated that higher average academic success of team
members yield more successful teams. Secondly, as range
of team members’ success increases, a decrease in team
performance is observed. Thirdly and finally, as GPA of
team members’ increase, their contribution to team
communication is increasing, which is an important
driver of success in this environment.
3. Methodology
3.1 Overall Design of the Study
In this research, two independent student team projects
which are conducted in fall and spring semesters of 2009
are used. Students from three universities, namely Atilim
University (AU), Universidad Tecnológica de Panamá
(UTP) and University of North Texas (UNT), are attended
to these projects. Scope of their works are based on design
and implementation of their assignments. In addition,
they are provided with a set of online communication
tools and they are trained on how to use them. Usage
statistics of these communication tools are collected and
students were pre-informed about this. Full coverage of
these projects and data collection were in the scope of the
article of Serçe et al. [7].
2. Relevant Research
2.1 Individual Success
Individual characteristics like academic success, age, sex,
education level and experience contribute on team
performance in collaborative teams [3]. In addition, some
researches indicate that with the increasing range of
individual characteristics within group leads to decrease
in team performance [5]. When the scope of individual
characteristics is limited to individual success, it can be
concluded that one person’s educational background and
work experience can considerably affect the perception on
his/her work [6].
3.2 Projects and Participants
As already mentioned, in this study, two projects which
are already undertaken by other researchers are used.
These projects were studied in the paper of Serçe et al.
with the name of “Online Collaboration: Collaborative
Behavior Patterns and Factors Affecting Globally
Distributed Team Performance” [7]. Important aspects of
these projects related to this study can be summarized as
following:
2.2 Globally Distributed Teams
Although some of the drawbacks of globally distributed
teams are mentioned, both industry and institutions
continue to work with small groups which are
combination of people from different regions [5].
Researches also mention that student teams and
professionals cannot be always analyzed with the same
reasoning, therefore the conclusions inferred from
researches based on student teams should not be followed
in industry without delicate analysis [3].
Project #1: In this project, student teams are assigned with
a database management system for car rental agency in
the fall semester of 2009 for 6 weeks. Scope of the work is
based on design, functionality assessment, implementing
and testing. Participants were from Atilim University
(AU), Universidad Tecnológica de Panamá (UTP) and
University of North Texas (UNT).
Project #2: In this project, student teams are assigned with
a standalone bookstore management application which
can be used by bookstore staff for daily operations. Scope
of this project was design and implementation of this
application and hence the duration was nearly two months.
Participating universities are same with the Project #1 and
this project is conducted during the spring semester of
2009.
2.3 Team Performance
Since the projects in this research based on capabilities
provided by online collaboration, effect of communication
on team performance should be considered. Research
about this issue reveals that frequency of communication
affects the performance of the team [3]. Because
communication frequency shows the number of times
team members interacted each other, the mentioned result
is not surprising but detailed analysis should also be
undertaken to check the relationships and underlying
effects.
Number of students with their universities and their roles
in these projects are summarized in Table 1 and Table 2
as following.
16
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Table 1: Number of students from each university in projects
AU
UNP
UTP
Project #1
38
29
12
Project #2
10
7
36
Table 2: Roles of students from each university in projects
AU
UNP
UTP
Project #1
Java Programmer
Tester / Leader
Database
designer
Project #2
Java Programmer
Tester
Database
designer
Fig. 1 GPA Histogram of Project #1
3.3 Measures and Data Collection
There are three important aspects of data related to this
study and they are GPA for indicator of individual success,
team performance grades for measuring team success
levels and communication statistics which are directly
influencer of team success in this environment [3].
Although vast amount of communication related statistics
are collected, there are significant level of missing data in
GPA values of students which are counted as missing and
removed from data completely in analysis stage. This loss
data restricted this research conducting comprehensive
statistical analysis and thus simple methods are used for
correlation and revealing underlying relationships.
Fig. 2 GPA Histogram of Project #2
Secondly, general situation of the team performances is
analyzed for both projects. While conducting these
projects, all students were assigned with a performance
grade and for each team, team averages are taken as team
performance grade in this study. With the same approach
used in the prior step, analysis of team performance
values are tabulated in Table 4. It can be seen that,
average performance grade and range is slightly shifted in
Project #2. In addition, when histograms of performance
grades are checked from Figure 3 and 4, it can be seen
that the second project grades are articulated and higher
than the first project grades.
3.4 Data Analysis
Firstly, in order to present the general situation, GPA
values of all participants are analyzed for each project. As
tabulated in Table 3 below, it can be seen that projects
have no significant difference in GPA values in the sense
of mean, standard deviation and ranges. In addition, when
their histograms are checked from Figure 1 and 2, even
there are less number of samples no significant skewness
in the graphs is observed.
Table 4: Analysis of performance grades for both projects
Average GPA
Std. Deviation
Min
Table 3: Analysis of GPA for both projects
Average GPA
Std. Deviation
Min
Max
Project #1
2,93
0,64
1,89
3,92
Project #2
2,55
0,68
1,25
3,86
Max
Project #1
64,75
14,41
50
91,50
Project #2
80,80
10,63
65
97,50
17
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Fig. 3 Performance Grade Histogram of Project #1
Fig. 5 Average Team Performance vs. Average Team GPA
After the analysis of the average team success, effect of
the most successful team player on the average team
success is analyzed. With this aim, for each team, student
with the highest GPA value is selected and relationship
between the average team performances is checked.
Although it is expected that having a team member with
higher GPA values can push team to have higher grades
in general, this situation does not hold for both projects.
In other words, where Project #1 has positive correlation,
Project #2 reveals a negative correlation between these
measures as it can be seen from Figure 6.
Fig. 4 Performance Grade Histogram of Project #2
After presenting the individual analysis of measures,
thirdly, effect of average GPA of development team on
team performance is analyzed. For each team, mean of
GPA values are calculated which is an indicator of the
general academic success of the team. With the same
approach used in the prior analysis, average performance
grades are calculated as team performance indicator.
Relationship between these measures can be seen from the
Figure 5 below. Scatter diagram shows that for both
projects, there is a tendency for increase in average team
performance as average team GPA increases. Since the
second project has higher performance grades in general,
effect of average GPA is not standing out as the first
project.
Fig. 6 Average Team Performance vs. Maximum Team GPA
As mentioned before, researches indicate difference in
team members’ individual characteristics yield decrease
in team performance. Considering the focus of this paper,
maximum difference in the GPA of team members for
each team is calculated. When the relationship between
GPA difference within team and average team
performance is checked from the Figure 7 below, negative
correlation can be easily seen. In other words, for both
projects as the GPA range within students increase
average team performance decreases.
18
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
between measures can easily be detected. Considering the
necessity for further studies, some conclusions from these
analysis are generalized and drawn:

When the average GPA of a team increases, team
performance is escalated. This shows that
individually successful team members can work in
harmony and create a successful teamwork
environment, which ultimately yields teams with
higher performance.

Increasing range of GPA decreases the team
performance. This confirms the researches
mentioning that the difference in individual
characteristics yields decreasing team success. This
conclusion is based on the fact that as range increases,
there is at least one team member with decreasing
GPA. Having unsuccessful students in the teams, it is
inevitable that the overall team performance will be
falling down.

Team members with higher GPA communicate more
in teams. This conclusion is based on the fact that, for
these projects each type of communication is counted
as same without considering their characteristics.
Since communication frequency shows the number of
interactions, as communication increases it can be
concluded that “a work is in progress”. This is
confirmed by the two projects, where in each
percentage of the team communication increases with
the higher GPA of the students.

Although team members with higher GPA values are
expected to pull their teams to be more successful, for
these two projects this statement cannot be verified in
this study. This is mostly based on the fact that, only
one individual with high GPA cannot affect the
whole team’s work attitude and performance grade.
This asymmetry in the teams resulted with the fact
that no concrete relation between the highest GPA in
the team and team performance can be found.
Fig. 7 Average Team Performance vs. Maximum GPA Difference in Team
In the relevant research section, it is mentioned that
studies reveal the fact that frequency of communication
affects the performance of the team [3]. Considering this,
how GPA of a team member is related to his/her
contribution to communication is analyzed. Before going
into further analysis, it should be mentioned that there are
many different communication tools available for these
projects, like forums, email, chat and wiki. However,
since analyzing and assessing differences and
contributions of each tool is out of scope of this paper, all
of them is counted as communication session. For each
student, percentage of their contribution to team
communication is calculated and its relationship to GPAs
are checked. From the scatter diagram in Figure 8, it can
be seen that there is a positive correlation between GPA
and contribution to communication for both projects.
Fig. 8 Team Communication Contribution vs. GPA
5. Conclusion and Future Work
In this paper, effect of individual success on globally
distributed software teams is studied. In order to reveal
underlying relationships, student team projects which are
conducted by other researchers and their data is used. In
these projects, students from Panama, Turkey and the
United States participated and their individual
characteristics and communication statistics are collected
4. Discussion
Data analysis showed that there are different underlying
relationship between GPA and team success. Although
there is high level of missing data, which restricts
undertaking detailed statistical analysis, correlation
19
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
conflict and performance in workgroups. Administrative
science quarterly, 44(4), 741-763.
[7] Serçe, F. C., Swigger, K., Alpaslan, F. N., Brazile, R.,
Dafoulas, G., & Lopez, V. (2011). Online collaboration:
Collaborative behavior patterns and factors affecting
globally distributed team performance. Computers in human
behavior, 27(1), 490-503.
[3]. Since data is collected and shared by other
researchers, there was a noteworthy level of missing data;
however, projects with the most complete dataset are
selected and used in this paper. Analysis of the data with
the goal of extracting underlying relationships between
the individual success of people and the successful
collaborations in globally distributed teams is undertaken.
In the light of the relevant research and this analysis,
some remarkable findings are constructed and presented.
Onur Yılmaz The author received B.Sc. degree from Industrial
Engineering department of Middle East Technical University (METU)
in 2012. In 2013, he gained second B.Sc. degree from Computer
Engineering department at the same university. Since 2012, he is a
master’s degree student in Computer Engineering field at METU. As
a professional, Onur Yılmaz works as software engineer and follows
opportunities in the internet industry.
For future work on this subject, there some important
points to mention. Firstly, in this paper a comprehensive
statistical analysis is not implemented due to low number
of data. However, with a complete and large dataset,
statistical analysis and validation should be undertaken to
clarify and support findings. Secondly, in this paper
student teams are studied but the conclusions provide
insight for implementing in the industry. Before
implementing these findings in the professional life,
delicate analysis of compatibility should be undertaken.
Thirdly and finally, dimension of individual success is
studied as GPA in this study. However, studies indicate
that individual success can be defined on multiple
parameters such as work experience level or knowledge
on the related subject. Considering this, different
parameters of individual success can be studied to reveal
their relationships to team performance. With the help of
these future studies, the presented conclusions and
analysis will be more concrete and drawbacks of them will
be overcome.
References
[1] Singh, I. (2011). Effectiveness of different personalities in
various roles in a software engineering team. Informally
published manuscript, Carnegie Mellon University,
Pittsburgh, PA, USA.
[2] Meyer, B. (2006). The unspoken revolution in software
engineering. Computer, 39(1), 124-121.
[3] Swigger, K., Nur Aplaslan, F., Lopez, V., Brazile, R.,
Dafoulas, G., & Serce, F. C. (2009, May). Structural factors
that affect global software development learning team
performance. In Proceedings of the special interest group on
management information system's 47th annual conference on
Computer personnel research (pp. 187-196). ACM.
[4] Dutoit, A. H., & Bruegge, B. (1998). Communication
metrics for software development. Software Engineering,
IEEE Transactions on, 24(8), 615-628.
[5] Bochner, S., & Hesketh, B. (1994). Power distance,
individualism/collectivism, and job-related attitudes in a
culturally diverse work group. Journal of Cross-Cultural
Psychology, 25(2), 233-257.
[6] Jehn, K. A., Northcraft, G. B., & Neale, M. A. (1999). Why
differences make a difference: A field study of diversity,
20
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
A Survey of Protocol Classifications and Presenting
a New Classification of Routing protocol in ad-hoc
1
Behnam Farhadi
, Vahid Moghiss2, Hossein Ranjbaran3, Farhadi.Behzad4
Scientific-Research Institute of Iran sensor Networks
Tehran, Iran
1
[email protected]
2
[email protected]
3
[email protected]
4
[email protected]
Abstract
Many Researchers are researching in the field of Wireless
Networks around the world. The results of these researches are
the annual production of several new protocols and standards.
Since the field is divided into several sub- categories so, each
new protocol is presented must place in this sub-categories based
on its own structure. Many researchers proposed new ideas in
the field of protocol classifications that they seem appropriate
categories in the last few years. According to the expand area of
wireless network research and studies on intelligent methods
right now and classifications are inefficient and need to be
designed a new classification. One of the most fundamental
problems in this area is the lack of a comprehensive structure
classification to accommodate the new intelligent protocol. In this
paper, we are going to compare different protocols and available
classifications, presenting a new classification by modeling from
Mendeleev table where protocols can be classified in one of the
branches of the tree, and allocate a unique code in the world.
problematic because of limited sources of these networks.
Using these protocols in routing made formation of loops. To
solve these problems, some protocols were designed for adhoc networks. Regarding the properties of ad-hoc networks,
due to variable structure the limitation of wireless
communication bandwidth, limited energy of nodes, each of
the designed protocols is confronted with problems. In this
paper, we review different unintelligent routing protocols in
mobile ad-hoc networks, which include routing algorithm
based on network structure in terms of on-demand and on
table-driven routing according to new classification. First, we
explain different protocols briefly and compare them in terms
of various parameters. The paper is organized as the follow: in
section II, we review related works and in part, (A) we analyze
unintelligent protocols based on network structure or on
routing table. In part, (B) we pay attention to unintelligent
protocols based on routing table and in part, (C) we pay
attention to unintelligent protocols based on network structure
in terms of hybrid methods. In section III, we introduce a
new method in classifying routing protocols. Conclusion and
further works are at the last section.
Keywords: Mobile ad-hoc network; Unintelligent routing
protocols; Modern classification.
1. INTRODUCTION
Mobile network is the network in which the relative
situation of its constituent nodes is varying. The routing
protocol must manage the node mobility and transfer the
parcel to target accurately so that communication sides is
not informed about the mobility in network nodes. One of
the mobile wireless networks, which it has been recently
paid attention [1], is ad-hoc network consisting of mobile
nodes while it does not use any control center in routing
its packets. These networks are applied in specific
situation in which centrality, wire support is impossible to
provide, or it is not economical to use ordinary systems
such as scientific conferences, military communications,
or saving the wounded in the natural disasters. By testing
the protocols used for wired networks, researchers
concluded that they are suitable for these networks .In
these protocols, due to the mobility in network, a lot of
routing overloading is created which it will be
2. REVIEWING PREVIOUS WORKS AND
OFFERING A COMPARATIVE TABLE FOR
EACH GROUP
2.1. Investigation and analysis of unintelligent protocols
based on network structure in terms of on-deman
Routing protocol of DSR [3] is a protocol in terms of ondemand that uses source routing in the algorithm that each
packet contains complete information of routing. Therefore,
each node is aware of information about its own neighbors. In
DSR protocol, each node uses the technology of route storing
to support route information that reduces routing overload.
This algorithm provides the shortest path available for the
enquirer.
21
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
TABLE 1. ON-DEMAND ROUTING PROTOCOLS
Network wide
broadcasting
N/A
rediscovery
low
Network wide
maybe
broadcasting
no
N/A
no
N/A
rediscovery
low
Network
Wide
broadcasting
low
Network wide
broadcasting
yes
no
yes
Shortest path
no
no
Participate in
routing
yes
no
yes
N/A
on demand
QOS
no
Update destination
high
Shortest and
fastest path
yes
Update period
Start repair at
failure
point
no
flat
Based on
occurrence
Source
node
AODV
[15]
on demand
flat
Based on
occurrence
Neighbor
nodes
TORA
[4,5]
on demand
flat
Periodically
& Based
on
occurrence
Source
node
ABR
[7]
periodically
Neighbor
nodes
ACOR
[7]
Based on
occurrence
Source
node
DSR
[3]
periodically
Neighbor
nodes
on demand
yes
Shortest path
no
no
on demand
(Reactive)
no
improvement of
convergence
no
no
on demand
Routing protocol of TORA [4,5] is an on-demand protocol
in terms of bi-directional links that increases the transmission
of packet numbers along the path thereby increasing the
overload in this protocol. TORA has found the ability to be
used in ad-hoc networks with high densities by modifying the
method of using bi-directional links.
Structure
no
yes
Route computation
maybe
N/A
Multicast capability
yes
N/A
Hello message requirement
yes
high
Route metric
Loop free
no
Unidirectional link
Broadcasting of route
request
Network
Wide
broadcasting
high
no
DV
N/A
Overhead
yes
Start repair at
failure
point
N/A
N/A N/A
yes
yes
Route recovery
yes
DV
Aware of distance between
node
yes
Usage Algorithm
Distributed
yes
flat
flat
flat
property
Protocol
DYMO
[11]
to solve the problem. Network lifetime increases by the
protocol that makes it suitable for large networks with high
density.
DYMO routing protocol [11] with a flat structure is the
main aim of which is to improve the convergence in network
that prevents from loop and count to infinity problems in
routing. DYMO protocol uses discover and storage processes.
With the breaking of each path, the discovery process is started
again. This protocol does not use periodical messages of
HELLO, which result in the reduction of routing overload
thereby meeting the requirements of high dense networks.
ABR routing protocol [7] is an on-demand protocol in
terms of source routing, which does not keep and store routing
information of nodes while founded or relatively stable paths.
(I.e. the paths are valid for a certain time in which nodes are
without movement). In this protocol, the chosen path is the
most stable although there is the possibility of unexpected
movement of nodes that the reconstruction trend of path will be
started of course, the path may not be reconstructed, and the
source is forced to start discovering from onset. This protocol
has a flat structure and uses HELLO messages periodically,
which increases overload. Therefore, this protocol is not
suitable for high dense and great mobile networks but it is
suggested for networks in which lifetime is a priority as it
produces paths of lifetime.
AODV Routing protocol [15] is a reactive protocol, which
needs the support of routing information about active paths.
This protocol uses some tables to restore the path information.
The number of route request exchanges of message and request
response is great. As the diffusion of these messages is
performed through distribution, there are extra messages
produced which reduces the node battery power. Bandwidth
and network efficiency is the reason if message redundancy. At
present, this protocol is the best and commonest.
ACOR routing protocol [8] is based on demand and uses a
flat structure. The main aim is to improve the quality of routing
that is using QOS parameters. It does not use network sources
in unnecessary occasions and starts to discover the path when it
is necessary. The response time to the path request is great in
this protocol but ACOR improves end-to-end delay parameter
2.2. analysis of structure based on unintelligent protocols
and routing table
DSDV routing protocol [2] is a table-driven protocol which
is not suitable for high dense and great mobile ad-hoc
networks due to creation of large overload in the situation of
22
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
CBRP is a Hybrid protocol [12] that uses meshing. In this
vast mobility in nodes , unreasonable use of network
protocol, each node must support its own neighbor table and
resources to keep un used information, of course , DSDV
contains the information required for all cluster members. This
protocol could reduce routing overload by using two different
protocol provides the fastest and shortest route while its biggest
partial and whole updating. The main aim of this protocol is to
problem is using network resources unnecessarily. The
remove the loop problem and count to infinity that most tableoverload is high in this protocol due to using route-finding
driven protocols suffer.
table and distributing Hello messages periodically. It is suitable
OLSR routing protocol [6] is a table-driven protocol that
for high dense networks because the network never fails to
has improved the classic bind state route –finding algorithm in
work.
which every node distributes all route information to its own
neighbors, while OLSR sends the information to a selected
3. A NEW METHOD TO CLASSIFY ROUTING
node named MPR that, reduces the number of updates and
PROTOCOLS
overload remarkably. This is ineffective for high dense and
vast movable ad-hoc networks due to dependency on routing
In this paper, we present a comprehensive classification for
tables.
routing protocol, which covers all-important items of expand
routing.
IARP routing protocol [9] is a table-driven protocol with
hierarchical structure that supports the routes immediately. It
does not periodic messages of Hello that helps to reduce the
overload while functioning on the shortest route and route
repair. The main aim of this protocol is to meet routing
requirement without any attention to network resources and
overload. It cannot be suitable for networks with limited energy
resources.
A new classification offered in [21], the important factors
are the main problems, which does not consider. For example,
learner automata based intelligent algorithm [23], which pays
attention to routing error tolerability, cannot be placed in it. In
[22], the classification is introduced with only one factor of
situation of services, which limited. In [24] a relatively more
comprehensive classification, which again does not consider
any place for intelligent algorithms [25], shows only a
classification for mobile models. There are also other
classifications in [16-20].
FSR routing protocol [13] is a table-driven protocol based
on link-state algorithm that exploits a mechanism similar to
fisheye’s that makes the nearest node and/or the most qualified
route in priority. Therefore, the accuracy of routing information
in this protocol is dependent on the distance to destination,
which could decrease the network overload due to information
exchange with nearest nodes frequently. FSR is a better
performance compared with other protocols with link-state
algorithms, as it does not try to obtain information about all
available nodes.
All these support the idea that researchers have sought to
find a classification to cover their preferred algorithm. In our
proposed method, we consider a tree the root of which is
routing. The root has two main elements being weighed 1 or 0
based on being intelligent or unintelligent respectively. One of
the important aims of designing and allocating code to its
elements is that the algorithm located at the subset of the last
leaf will have a specific code after being surveyed and the
order of placing each algorithm is based on the publishing year
and the importance while the algorithm codes start from 1.
TBRPF routing protocol [14] is a table-driven protocol with
link-state algorithm, which provides the shortest route by hopby-hop routing method. Each node has a table, which keeps the
information of all accessible nodes. It uses periodic messages
of Hello to discover and supervise the route but the number of
Hello messages is fewer than usual. It aims at successful
routing and uses the network suitably, which makes it usable in
ad-hoc networks.
The left side element (unintelligent) is consisted of two
position-based and topology-based subsets position-based part
has two types of subset of aware and unaware parts. The
topology-based part is divided in to Topology-aware and
Topology-unaware parts. We have classified the former part in
to flat and hierarchical parts and the latter in to four distinct
parts. The right side element (intelligent) contains four subsets,
which the researchers have used to make the algorithms
intelligible. The subset of these parts is a part, which specifies
the main feature of intelligent algorithm and contains various
subsets dependent on algorithm. Take, for example, ZRP and
ZHLS algorithms. In our method, as in figure (1), the code of
algorithm ZRP is ( 001101 ) which lets us know that it is a
routing algorithm of unintelligent (the first digit 0), structure
based (second digit 0) and hierarchical (third digit 1) which has
hybrid method. Using the last digit, we know that it is the first
protocol in this class our proposed classification is shown in
figure (1). As shown in Fig (2), it is illustrated routing protocol
classification with new method.
2.3. analysis of Hybrid structure based on unintelligent
protocols
ZRP routing protocol [10] is a Hybrid protocol that uses ondemand method in intra zone routing and
table –driven
methods in inter zone routing. This will reduce control in tabledriven method and recession (The response time of route
request) in on-demand method. The network is divided in to
routing zones according to the distance between the mobile
nodes. ZPR protocol has considered different routing methods
to find the routes inter zone and intra zone that can provide the
shortest and best route. In addition, in this protocol the
overload is low which makes it suitable for high dense network
in which routing quality is in priority.
23
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
TABLE 2. TABLE-DRIVEN ROUTING PROTOCOLS
no
no
Table
driven
Update destination
Shortest path
Update period
Route computation
no
flat
hybrid
Neigh
bor
nodes
DSDV
[2]
MPR
node
OLSR
[6]
Structure
Multicast capability
no
Hello message requirement
N/A
Route metric
high
Unidirectional link
No
recover
Loop free
Broadcasting of route
request
yes
no
Overhead
NO
DV
Route recovery
yes
Aware of distance between
node
yes
Usage Algorithm
Distributed
yes
LS
no
Start
repair at
failure
point
high
Only
broadcasting
to MPRs
yes
no
Shortest path
yes
yes
Table
driven
flat
Periodically
LS
yes
Rediscov
ery
low
flooded into
the network
aybe
no
m Shortest path
no
no
Table
driven
flat
Periodically
LS
no
N/A
medium
yes
N/A
Shortest path
yes
yes
Table
driven
flat
Periodically
LS
no
N/A
medium
N/A
no
Shortest path
and route repair
no
N/A
Table
driven
hierarc
hically
Periodically
TABLE 3.
Update destination
Update period
Structure
Route computation
Multicast
capability
Hello message
requirement
Route metric
Unidirectional link
Loop free
Broadcasting of
route request
Overhead
Route recovery
Aware of distance
between node
Usage Algorithm
Distributed
N/A
Periodically
flat
Hybrid
no
yes
ZRP [10]
Property
IARP
[9]
1
intellig
Unintelligent
0
1
Hybrid
Topology Base (Network Structure)
no
yes
yes
Network Wide
broadcasting
NO
TBRPF
[14]
Routing
Periodically
hierarchical
yes
Only broadcasting to
clusterheads
LS
FSR
[13]
N/A
0
Shortest path
yes
Neigh
bor
nodes
Neigh
bor
nodes
Neighbor nodes
Shortest and fastest
path
yes
high
Start repair at failure
point
Protocol
with a unique code. Our future work is to develop the proposed
tree in all areas of ad-hoc networks such as security, secure
routing, coverage, and research about intelligent algorithms.
HYBRID ROUTING PROTOCOLS
CBRP [12]
Protocol
Network
Wide
broadcast
Broadcast
To
Intrazone
Property
0
1
Flat
no
Position Base
hierarchical
0
Method
medium
no
DV
0
Yes
On-Demand
(reactive)
N/A
2
1
Table Driven
(ProActive)
Hybrid
4. CONCLUSION AND FURTHER WORKS
1-ZRP
Routing is an important component of communication
protocols in mobile ad-hoc networks. The design of a protocol
in these networks is performed to obtain special objectives,
based on available requirements, network specifications, or
application in different zones. In this paper, we introduce
variety of routing protocols along with its applied features. We
present a new classification (taxonomy) method, based on
Mendeleev table in which each algorithm with a new code.
2-ZHLS
Figure 1. Example of our presented classification
This tree is undergoing the development and each
researcher can locate his or her algorithm in one of the classes
24
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Fault Tolerance
Ant Colony
decrease of power
Learning
Automata
location finding
primary object
Intelligent
Quality
Neural Network
data delivery
Genetic
Energy aware
Flooding, Gossiping And
controlled flooding
Data-centric
Network topology
unaware(local
information)
Routing
Energy - aware
qos routing
topology
based(network
structure
Cluster- based
table driven
(proactive)
link state
hop-count
hierachical
unintelligent
Network topology
aware(global
information)
method
hybrid
qos routing
flat
ondemand(reactive)
link state
hop-based
unknown(virtual)
cordinate
hop-count
distance based
position base
routing loop
recovery
guaranted
delivery
routing loop
avoiding
known(physical)
coordinate
best effort
delivery
Figure 2. A view of routing protocol classification with new method
25
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
[14] R. Ogier, "Topology Dissemination Based on Reverse-Path Forwarding
(TBRPF) : Correctness and Simulation Evaluation", Technical Report,
SRI International, Arlington, Va, USA, Oct. 2003.
[15] C.E. Perkins, E.M. Royer, "Ad Hoc On-Demand Distance Vector
Routing." Proceedings of the 2nd IEEE Workshop on Mobile
Computing Systems and Applications (WMCSA ’99), pp. 90-100, New
Orleans, LA, Feb. 1999.
[16] Miguel A. Labrador, Pedro M. Wightman, ―Topology Control in
Wireless Sensor Networks,ǁ Library of Congress Control Number:
2008941077, Springer Science + Business Media B.V, 2009.
[17] A. Nayak, I. Stojmenovic ―Wireless Sensor and Actuator Networks
Algorithms and
Protocols for Scalable Coordination and Data
Communication,ǁ Published by John Wiley & Sons, Inc., Hoboken, New
Jersey ,Canada, 2010.
[18] A. Boukerche, ―Algorithms and Protocols for Wireless Sensor
Networks,ǁ Published by John Wiley & Sons, Inc (Wiley-IEEE Press),
Nov. 2008.
[19] J. Zheng , A. Jamalipour ―Wireless Sensor Networks: A Networking
Perspective,ǁ Published by John Wiley & Sons, Inc (Wiley-IEEE Press),
Oct. 2009.
[20] K. Sohraby, D. Minoli, T. Znati ―Wireless Sensor Networks:
Technology, Protocols, and Applications,ǁ Published by Wiley-IEEE
Press, May 2007.
[21] A. N. Al-Khwildi, K. K. Loo and H. S. Al-Raweshidy ―An Efficient
Routing Protocol for Wireless Ad Hoc Networks,ǁ 4th International
Conference: Sciences of Electronic,Technologies of Information and
Telecommunications (SETIT 2007), Tunisia, Mar. 2007.
[22] S.M. Das, H. Pucha, and Y.C. Hu, "Performance Comparison of
Scalable Location Services for Geographic Ad hoc Routing," In Proc. of
the IEEE Infocom'05, vol. 2, pp. 1228-1239. Miami, FL, Mar. 2005.
[23] S.M. Abolhasani, M. Esnaashari, M. R. Meybodi, "A Fault Tolerant
Routing Algorithm Based on Learning Automata for Data Inconsistency
Error in Sensor Networks," Proceedings of
the 15th Annual CSI
Computer Conference (CSICC'10), Tehran, Iran, Feb. 2010.
[24] Krishna Gorantala "Routing Protocols in Mobile Ad-hoc Networks,"
Master Thesis, Department of Computing Science, Ume°a University,
Ume°a, Sweden, Jun. 2006.
[25] H. Idoudi, M. Molnar, A. Belghith, B. Cousin "Modeling Uncertainties
in Proactive Routing Protocols for Ad Hoc Networks," Information and
Communications Technology'07 (ICICT 2007), ITI 5th IEEE
International Conference, pp. 327-332, Dec. 2007.
REFERENCES
[1]
[2]
[3]
[4]
[5]
[6]
[7]
[8]
[9]
[10]
[11]
[12]
[13]
Murthy, S. and J.J. Garcia-Luna-Aceves, ― An Efficient Routing Protocol
for Wireless Networks,ǁ ACM Mobile Networks and App. J., (Special
Issue on Routing in Mobile Communication Networks), pp. 183-97, Oct.
1996.
C. E. Perkins, P. Bhagwat, ―Highly Dynamic Destination-Sequenced
Distance-Vector Routing (DSDV) for Mobile Computers,ǁ ACM
Computer Communication Review (ACM SIGCOMM’94), vol. 24, no.
4, pp. 234-244, Oct. 1994.
D. Johnson, D. A. Maltz, ―Dynamic Source Routing in Ad Hoc Wireless
Networks,ǁ in Mobile Computing (T. Imielinski and H. Korth, eds.),
Kluwer Acad. Publ., 1996.
V. D. Park, M. S. Corson, ―A Highly Adaptive Distributed Routing
Algorithm for Mobile Wireless Networks,ǁ INFOCOM ’97, Sixteenth
Annual Joint Conference of the IEEE Computer and Communications
Societies. Driving the Information Revolution, Proceedings IEEE, vol. 3,
pp. 1405 -1413, Apr. 1997.
V. D. Park, M. S. Corson, ―Temporally-Ordered Routing Algorithm
(TORA) Version 1 Functional Specification,ǁ IETF Internet Draft, 1997.
T. Clausen, P. Jacquet, "RFC 3626 - Optimized Link State Routing
Protocol (OLSR)," Project Hipercom, INRIA, Oct. 2003.
C.-K. Toh, ―Associativity Based Routing For Ad Hoc Mobile
Networks,ǁ Wireless Personal Communications Journal, Special Issue
on Mobile Networking and Computing Systems, vol. 4, pp. 103-139,
Mar. 1997.
Jinyna Li, J. Jannotti, D. S. J. De Couto, D. Karger and R. Morris, ―A
Scalable Location Service for Geographic Ad Hoc Routing,ǁ in Proc.
MOBICOM’2000, Boston, MA, USA, 2000.
Zygmunt J. Haas, Marc R. Pearlman, Prince Samar, ―The Intrazone
Routing Protocol (IARP) for Ad Hoc Networks,ǁ IETF Internet Draft,
Jul. 2002.
Z. J. Haas, M.R Pearlman, ―The Zone Routing Protocol (ZRP) for Ad
Hoc Networks,ǁ IETF Internet Draft, Aug. 1998.
I. Chakeres CenGen, C. Perkins, WiChorus, ―Dynamic MANET OnDemand (DYMO) Routing,ǁ IETF Internet Draft , Sept. 2009.
Mingliang Jiang, Jinyang Li and Y. C. Tay, ―Cluster Based Routing
Protocol (CBRP),ǁ IETF Internet Draft , Aug. 1999.
G. Pei, M. Gerla and T.-W. Chen, ―Fisheye State Routing in Mobile Ad
Hoc Networks,ǁ In Proceedings of the 2000 ICDCS Workshops, pp.
D71-D78, Taipei, Taiwan, Apr. 2000.
26
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Developing mobile educational apps: development strategies,
tools and business models
Serena Pastore
1
Astronomical Observatory of Padova, INAF
Padova, 35122, ITALY
[email protected]
with the several market analysis companies [2]. Mobile
actors involved in the mobile ecosystem (Fig. 1) includes
devices manufacturers (e.g., Nokia, Samsung, RIM
Blackberry), software platforms providers (e.g. Apple;
Google, Microsoft), applications developers and mobile
operators that in each country (e.g., in Italy Vodafone,
TIM, 3G) provide Internet connectivity. A single mobile
device is composed by specific hardware and software
platform that deals with the peculiarities of these systems.
Mobile devices should follow the user's mobility
(therefore mobile devices are generally of reduced size
and weight, such that they can be transported easily).
Their hardware components manifest limited processing
power and memory, little screen dimension, but a variable
numbers of other hardware such as cameras, network
adaptors (Wi-Fi, Bluetooth), sensors like GPS (Global
Positioning systems) receivers, accelerometer, compass,
proximity sensor, ambient light sensor (ALS), and gyros.
Abstract
The mobile world is a growing and evolving market in all its
aspects from hardware, networks, operating systems and
applications. Mobile applications or apps are becoming the new
frontier of software development, since actual digital users use
mobile devices as an everyday object. Therefore also in an
educational environment, apps could be adopted in order to take
advantage of mobile devices diffusions. Developing an app
should not be a decision linked to the trends, but must follow
specific strategies in choosing target devices, mobile platforms,
and distribution channels that in these contexts usually go
through e-commerce sites called stores (e.g., Apple Store,
Google Play, Windows Phone store). Furthermore the design of
an educational mobile app requires a careful analysis on the
methodologies to be adopted and the available tools especially if
the aim is to build a successful and useful app. Very often there
is a tradeoff between the costs associated with the development
that requires a lot of technical and programming skills. The
economic return is neither high nor guaranteed considering the
business models offered by the major stores. The paper deals
with the approaches to apps development analyzing
methodologies, distribution channels and business models in
order to define the key points and strategies to be adopted in an
educational context.
Keywords: Mobile apps development, cross-platforms mobile
tools, native apps, e-commerce, hybrid apps, Sencha Touch,
Apache Cordova
1. Introduction
The mobile world is a growing and evolving market
in every its facets. Mobile devices [1] intended as
electronic devices created for specific functions (e.g.,
voice like cell phones and smartphones, reading books as
e-book reader, and more general activity such as laptops
or tablets) have undergone exponential growth in recent
years, especially with the introduction in the market of
mobile devices connected to the Internet network as
smartphones and tablets. Even if such market includes the
different categories of mobile devices, tablets and
smartphones seem to be the top categories in accordance
Fig. 1 Actors of the mobile ecosystem
Software applications on such devices or mobile apps as
they are universally called, are the new frontier of
software considering the mobile devices diffusion and
should follow the rules of such ecosystem. Developing an
app of every category, even in an educational environment,
should be the result of a careful analysis of mobile
platforms, tools and methodologies to be adopted. There
27
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
are a lot of hardware manufactures (e.g., Apple, Samsung,
Nokia, LG, HTG, Asus) and this fragmentation also
breads in mobile software platforms (e.g., iOS, Android,
Symbian, and Windows Phone) that directly deal with
hardware and help to take advantage of the device's
performance. An app is usually targeted to few software
platforms, since developing for all existing platforms is
really expensive and requires high technical skills. The
paper deals with the approaches to apps development
focusing on methodologies, tools and distribution
channels to analyze peculiarities and issues of the
different methods. From methodologies point of view,
apps are divided into native and web apps [3] meaning the
use of different technologies. Moreover a compromise
between the pro and cons of the two methods are the
hybrid apps that use special frameworks to convert web
apps into native apps. There are different aspects that
should be taken into consideration when choosing a
specific platform or methodology, and the paper would
like to analyze them in order to figure out which is the
best approach for educational apps. The first section
analyzes apps’ features focusing on types and the
structure of their market. The second section distinguishes
the choices in the development methodologies and tools,
while the third section examines the tools needed for the
development of each kind of app. It will finally give an
evaluation of the various types of development taking into
account also the distribution channels of the store
normally used by users to search for an app.
mobile platform and are therefore capable of interfacing
completely with the hardware’s device. But since they are
dedicated to mobile devices equipped with specific
operating system, they are targeted to a single platform.
Web apps use web technologies (HTML5, JavaScript,
CSS3 [4]) and are multi-platform, since they can be
executed from any mobile device equipped with a web
browser. However such app could not take advantage of
all hardware capabilities, since interacts with the
hardware by means of an intermediate software layer and
specific Application Programming Interfaces (APIs). A
compromise between the pro and cons of the two methods
are the hybrid apps that use special frameworks mainly
based on JavaScript technology to convert web apps into
native apps [5].
Usually apps could be preinstalled in the device and
thus imposed by the mobile software reference platform or
could be installed by the user or, as in the case of web app,
imply a request to an Internet server machine. Convince a
user to download and use an app depends a lot on both its
utility and its quality and how it is distributed.
As regards apps distribution, even if some app can be
delivered through specific websites or e-mail, the main
distribution channel is an e-commerce site called store [6]
owned by major hardware and software mobile providers.
Each store imposes specific policies for apps distribution
and different business models. In this context, a developer
should decide if developing for one or more mobile
platforms, since this decision affects the choices on
methodologies, technologies, tools and distribution
channels.
2. Mobile apps features and the business
models
2.1 Mobile devices and platforms
Each mobile device has different technical features in
terms of available memory, computing power, screen
resolution,
operating
system
(OS)
and
the
hardware/software platform that determine the quality and
performance. Even if actually the market seems to be
occupied by few dominant players (Apple and Samsung),
there are many other manufacturers of hardware and
software that are trying to earn their slice market. Many
companies producing mainly software such as Microsoft
and Google have made agreements with hardware
manufacturers to provide their own systems. Other
companies like Apple or RIM Blackberry provide both the
hardware and the software. Each manufacturer tries to
keep pace with new technologies, thereby spreading out
ever newer models of the two main categories of devices
(i.e., smartphones and tablets) trying to anticipate
competitors to increase their market value.
Among the leading manufacturers of mobile
hardware, Nokia has been at the top of the sale of mobile
App design (Fig. 2), starting from the concept idea
and requirement analysis, mainly differs in the stages
concerning the development and the final delivery.
Fig. 2 Apps’ development and deployment steps
As regards development, apps can be divided into
native or web apps meaning the use of different
programming technologies and thus different tools and
frameworks. Native apps use the languages of the specific
28
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
phones for many years, offering devices running the
SymbianOS operating system. But with the arrival of
smartphones, Nokia has lagged behind its competitors
Apple and Samsung. It currently offers mobile phones
that still have the Symbian OS operating system, but also
the smartphones Lumia series that for now, thanks to an
agreement with the supplier of Microsoft software are
equipped with the Microsoft operating system Windows
Phone. Samsung has been for many years very successful
in selling low-end phones and smartphones. Samsung
uses as software platforms the BadaOS of his property
(even if it will be probably replaced with the TizenOS [5],
an open operating system developed in collaboration with
Intel) in most of mobile devices. It uses the Android
operating system in the smartphone Galaxy series which
are those that have decreed its success.
Sony-Ericsson has equipped its devices with
Windows Phone and Android in the latest Xperia
products, while Research in Motion (RIM) is linked to its
successful BlackBerry product, now equipped with the
operating BlackBerry OS system version 10 as other
mobile devices provided by such company. LG instead
equips its mobile devices with Android.
iOS and Android actually are predominant in the
market of mobile OS, but other OS (e.g., Windows Phone,
Blackberry [5]) could, however, have their market and
could be the target platform for a developer. An app
provider very often has to optimize the software for a
specific platform or device. In order to obtain the best
performance and thus to maximize the mobile experience,
he/she should know the devices’ various mobile platforms.
An app depends strongly on the mobile context, then on
the type of hardware device, the software platform and the
network used for data transmission. Then apps final
delivery is usually done by means of stores. The user
searches on a store the app and, when selected, it is
automatically installed on the device. There is no way to
test the app before downloading and this makes it difficult
for the user to select the application on the basis of the
available information. Considering the spread of such
devices, Fig. 3 displays mobile users’ main activities.
Fig. 3 Time spent on iOS and Android device (April 2013)
You see as the percentage of time spent on iOS and
Android devices in the use of apps is greater than the time
spent in web browsing. This clearly indicates the
importance of an app.
Very often the factors that influence a user in
downloading an app, apart from the choice of the category,
are the price, a clear description of the purpose of the
sample images, and the vote given by the users. Rarely a
developer can affect the price in an important way,
considering the fact that users prefer to download free
apps, but it can affect the other elements. The description
and pictures showing the app in action are very important,
since they are the way in which a user can understand
how an app behaves. Once the user has downloaded the
application, he/she will expect that the app responds to
his/her needs. A simple and intuitive application is the
best method to be appreciated by a user. If the apps is
delivered by the reference store, there are other factors
that many influence the choice. Each store in fact
implements an evaluation method based on a vote that the
user needs to do. A developer must keep in mind that app’
future updates should be able to improve the user
experience and remedy the possible negative criticism.
Focusing on store distribution channel, each store
offers different business models to take a revenue from
apps, and studies demonstrate that this is the preferred
channel to search for an app. Recent study by Canalys
analyst firm on the main store (Apple's AppStore, Google
Play, Windows Phone and Blackberry Store World) [7]
shows a constant growth in apps sales during the first
quarter of 2013. Between the stores, Apple store records
the highest turnover of three-quarters of the total value of
the market (74%), while Google Play is the first
framework for number of app downloads (51%). The
other stores are more detached, but still remain the point
of reference for users and developers of related ecosystems.
29
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
2.2 Apps business models on mobile stores
On a store, app are usually downloaded according a
license that is set by the developers. However each store,
as the Table 1 shows, offers a specific program to sell the
app ranging from free/paid app for download to the others
business models (e.g., in-app purchase and in-app
advertising). The first distinction is about free apps and
paid apps. Then as regards the ways to monetize an app,
the two main models are pay per download or premium
offering and in app purchases (IAP). Pay for download
method was the first model adopted especially in the App
Store with different values of fees required (e.g., most
apps requires a little fee of $0.99 or $1.99). Each user
pays the app before downloading and installing into the
device. The in-app purchases (IAPs) is a model that lets a
user to download an app both for free (in this case the
model is known as freemium) or paid, and requires a
payment if the user want to enable some added features or
functionalities, make some upgrades or have game
advancement (e.g., enabling levels in a game). Another
source of revenue could be the in-app advertising or inapp ads.
App Store
Google
Play
Window
Free (no charge to
download)
and
priced (user charged
before download)
Free and paid app
whose base price
can be selected by
country
Blackberry
World
Free
and
paid
(purchase
before
downloading)
apps and apps divided into
consumable or durable);
- ad-funded apps (free
apps with ads made by
inserted code thanks to
Microsoft
Advertising
service
-7-day,
30-day
subscription (trial period
after than users pay a fee
to
renew
their
subscription)
Other analysis however have shown (e.g., by Distimo
[8]) that the in-app purchases in the fremium model
generate the majority of revenue in the app stores. Finally
the in-app advertising model uses the store advertising
platform or service and consists in inserting specific
advertising lines to the app code in order to take a revenue
from such publicity. According to recent statistics ads
through mobile or mobile advertising is increasing, even
if this model should necessary be different from the
traditional model of ads that dominates the Web sites. In
many cases, getting an advertising message when
download an app for free, seems to be a price that users
are willing to pay for having no costs. As the Table 1
describes, there are different models of advertising such as
affiliate marketing that promotes other mobile apps in
order to earn commissions when the app is purchased and
downloaded.
In any case the best strategy for marketing and app
success (Fig. 4) depends on several factors, including the
buying habits of the target audience and the type of
application. For example according several analysts,
Apple users are more used to pay for their apps and so
paid apps are normally. In other contexts, the fremium
model seems to prevail, since the app is offered for free
and you only pay for enhanced functionalities. However,
considering the number of free apps and the low revenue
from advertising network, most developer use other
channels to make money from creating mobile software.
Table 1: Business models for apps selling in the different stores
Fixed tires (e.g. Pay
for download)
Fixed tires: free,
$0.99, $1.99, $2.99,
… Exchange rate
fixed and controlled
by Apple. Price can
be changed anytime
Fremium
model
generates 72% of
total revenues of
paid apps
Phone
Store
Additional models
- In app purchases (offer
additional digital content,
functionality, services or
subscription within the
paid or free apps);
- iAd rich media ads
(exploiting the Apple’s
digital
advertising
platform and adds an ass
to receive 70% of the net
ad revenue generated)
volume
purchase
program
(a
specific
program for business and
education institution to
purchase apps in volume)
-in app products and
subscriptions (e.g., onetime purchases and autorenewing
subscription
from inside the app,
freemium);
- ad-supported model (ads
apps with the use of
AdMob integration or
distribution for free and
selling in-app products or
advertising)
- in-app product (e.g., trial
Fig. 4 Good practice for app design, marketing vs. milestones for app success
A recent report [9] shows as the major revenue comes
from other sources that app store sales and advertising
such as e-commerce licensing and commissioned appmaking. Probably the top revenue-generating apps on the
30
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
store belong to the game category or in some cases to
some mobile messaging app.
3. Strategies on apps development
App development project is not a simple action and
requires critical decisions on strategies to be adopted
considering the scope and the target. Usually apps
manifest some peculiarities: even if developed with the
same techniques used by other type of application
software, they should be light, essential and easy to use
considering the features of mobile devices. Even if a
mobile app could be distributed via a website or e-mail,
the main channel is the reference store of the mobile
software platform. Since each store is proprietary, it
imposes rigid policies on content and quality of the app
influencing its development. Furthermore there is a cost
variable to each store related to the store’s developer
program registration.
The focus is so on the mobile players and on the different
aspects of app’s development that range from the scope,
the software tools needed and the business model.
2.3 Apps category
There are categories of apps that seems to be more
appealing in a mobile context. For example the research
company Nielsen describes as in Brazil and in Italy, users
prefer to surf the Internet with the mobile device rather
than downloading apps, while in countries like Australia,
India, South Korea and Turkey the use of apps and web
browsing is comparable. Mobile apps and websites for
mobile devices should be different things referring to two
different activities, but sometimes many apps are
developed as versions of mobile optimized websites or just
links to websites. In any case, mobile websites are
designed to work on any mobile browser, while generally
the apps are designed for one or a few mobile platforms.
Focusing on apps, statistics on major stores shows as the
games represent a great percentage of used apps, even if,
as the Fig. 5 shows, maps (Google Maps) and social
network apps are in the top positions (Facebook,
YouTube, Google+, WeChat, Twitter) followed by
messaging apps (e.g. Skype, Whatsapp).
3.1 Tools needed for mobile apps development
Considering the development methods, the choice of
developing a native, web or hybrid apps includes the
choice of the tools and the frameworks. Table 2
summarizes the different tools available for native, web
and hybrid apps using several mobile platforms.
Table 2: Tools and frameworks for mobile development
Mobile
software
platform
iOS
Fig. 5 Statistics on the activities performed by users of smartphones at least
once a month in various countries (source: Nielsen survey, February 2013)
In some way software platforms discriminates also the
apps categories considering that mobile apps are
preinstalled on the device. If a developer needs to mediate
between the actual usages of the app and its cost, these
analysis are very important.
In the Italian context, for example, a report about mobile
users behavior [10] shows as mobile users have on
average 27 apps installed on the device, but about half use
every month and every day. Of these most of them are free
apps, and only 6 % of users have mainly paid app.
Android
Windows
Development tools for native, web and hybrid
app
App lifetime cycle: Develop, test, distribute
maintain;
Native app develop tools: Xcode (Knowledge of
Objective-C).
Native .IPA file to deploy on the device or on the
store;
Web app develop tools: Web frameworks
Javascript libraries (e.g., jQuery mobile, Zepto)
Knowledge of HTML5, CCS3, Javascript.
Hybrid app tools: Phonegap, Titanium
AppCelerator, RhoMobile suite by Motorola and
Rhodes open source Ruby-based framework to
build native apps, eMobc
App lifetime cycle: Setup, develop, debug&test,
distribute
Native app develop tools: Android development
kit (ADT) or Android Studio (knowledge of
Java) – packaged as apk file
Web app develop tools: Web frameworks and
Javascript libraries.
Hybrid app tools: Phonegap, Titanium Appc,
RhoMobile suite, eMobc
App lifetime cycle: design, develop, distribute
31
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Phone
Blackberry
OS
Javascript libraries built-in for mobile web apps (e.g.,
jQuery mobile, Zepto.js). Focusing on mobile framework,
some of this framework are simple JavaScript libraries
offering the developer a set of useful built-in functions.
Native app develop tools: Windows Phone SDK
(along with Visual Studio)
Knowledge of C#, C++, Visual Basic. Packaged
as XAP file.
Web app develop tools: Web frameworks and
Javascript libraries.
Hybrid app tools: Phonegap, Titanium
AppCelerator, RhoMobile suite
App lifetime cycle: design, develop, distribute
Native app develop tools: Momentic IDE
(Blackberry 10 Native SDK)
Web app develop tools: Web frameworks and
Javascript libraries.
Hybrid app tools: Phonegap, Titanium
AppCelerator, RhoMobile suite
Table 3: Tools and frameworks for mobile development
Framework
HTML5
framework
For native apps development, each mobile software
platform provides a custom software development kit
(SDK) that consists in a set of tools allowing you to
design, develop, build and manage the process of app
development. Most of them enable you to create a package
version of the app suited to be published on the target
store. Generally in each SDK there are compilers to
translate the source code in the mobile platform reference
language (e.g., Objective-C for iOS, Java for Android, C#
or C++ for Windows Phone, C/C++/Qt for Blackberry 10)
into an executable, the standard libraries with public
interfaces and other tools such as the visual editors and
the simulators in order to help the development. The
development environment takes advantages of the mobile
operating system and the software component related to
runtime environments (e.g., the Android Dalvik Virtual
machine or the Windows WinRT runtime environment)
and the applications frameworks (e.g., Windows
Silverlight, Apple Cocoa Touch). Mobile native apps are
designed for a target mobile platform by using native
programming languages and because of that they have
access to all features of the device ensuring optimum
performance.
Normally they require the installation into the device, a
local execution and not necessary they need an Internet
connection (except for social network or messaging apps).
The main distribution channel is the store.
Mobile web app are developed in accordance with the
HTML5 framework that understand the language of style
CSS3 and Javascript programming language implemented
using different dialects. There are different HTML5,
CSS3 and Javascript frameworks that simplify the
development process and speed up the coding.
Table 3 shows an example list of frameworks that could
be used to develop web apps. We tried to made a
distinction between tools that are used to create web
applications for desktop (i.e., 52 framework) from those
optimized for a mobile environment (e.g., Zoey), from the
JS
mobile
libraries
HTML5CSS3 mobile
framework
Software available
Iio Engine -Open source lightweight
framework for creating HTML5 applications
with JS and canvas. SDK + debugging system
+ cross-platform deployment engine
LimeJS -HTML5 game framework for
building games that work on touch screens
and desktop browser. Created with the
Closure library by Google.
52
framework -HTML5-CSS3
based
framework
Kendo UI – inclusive HTML5/JS framework
for modern web and mobile app development
Zepto.js – jS framework for mobile webkit
browser compatible iQuery
jQuery mobile – touch optimized web
framework built on top of jQuery
M-project – mobile HTML5 JS framework to
build mobile apps easy and fast
Xui – a super micro tiny DOM library for
authoring html5 mobile web apps
Sencha Touch – HTML5 mobile app to
develop app that look and feel native on iOS
and Android
Zoey -Lightweight framework for developing
mobile apps Built on Zepto.js, support modern
browsers (iOS 4.x and Android 4.x)
Jo -Open source mobile application
framework working on iOS, Android,
Symbian, Safari, Chrome and dashboard
widgets. Compatible with PhoneGap
Lungo.js -Mobile framework that includes
features of HTML5, CSS3 and Javascript
(iOS,
Android,
Blackberyy,
WebOs).
Distribution also on stores
Junior Front-end framework for building
HTML5 mobile apps that look and behave
native (uses Zepto.js and integration with
backbone.js views + routers)
eMobc – open source framework for
generation of web, mobile and native iOS and
Android apps using XML
The inclusion in the source code of the app is simple and
consists, as the Fig. 6 shows for the case of the Zoey
framework, in putting in the scripts and the built-in style
sheets inside the specific tags of the HTML5 code (e.g.,
the script and the link tags).
32
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
3.1.1 Sencha Touch
Sencha Touch is a Model View Controller (MVC) JS
framework that appears to be more complex than a single
library as seen before. Its use requires to download the
Sencha Touch SDK and Sencha Cmd that install other
software like Ant, Ruby, Sass and Compass. Sencha cmd
is necessary to generate the app, since it gives a set of
command line tools that allows to create the skeleton for
the application available through a web address. The
skeleton is composed by some directories (e.g., app that
contains the models, views, controllers and stores) and
files useful for the app. The app.js file is the entry point of
the app. It contains the launch function where the
application logic is inserted. The app is so made by an
index.html file and the Javascript file that contains the
source code of the app. Sencha touch comes with a set of
components, objects, functions and code ready to use. A
Sencha touch app is web-based, but the framework offers
a way to include JS commands (by using native API with
Ext.device function) in order to be translated into native
functions (e.g., orientation, notification) when the app is
compiled. The options to be set in order to transform the
app into native are configured in the packager.json file.
An example of the structure of a Sencha touch app is
shown in Fig. 8.
Fig. 6 Tools and frameworks for mobile development
The same figure shows an example of a web page coded
as single o multi-web page that uses the JQuery mobile
framework by including the Javascript library (jquerymobile.js) and the style sheet (jquery-mobile.css) in the
header section of the web page. Once included, the
various functions coded in the library and the instructions
inside the style sheet could be called directly in the body
of the page. Other frameworks are more complex and also
include a developing environment and the command line
tools. Moreover they could be targeted to some platforms
or development languages as the Fig. 7 shows.
Fig. 7 A comparison between mobile frameworks and
functionalities.
Fig. 8 Tools and frameworks for mobile development
For example Sencha company offers a set of products that
allow to design (Sencha Architect), develop (Sencha Ext
JS, Sencha GXT, and Sencha Touch) and deploy (Sencha
Space) apps for mobile and desktop. Of these products,
Sencha releases some of these software with an open
source license as the case of Sencha Touch that is the
framework dedicated to develop HTML5 Mobile app for
iOS, Android and Blackberry. Since we focus on an open
source framework that could allow to develop a mobile
app, we examine, from the technical skill required, three
products that are Sencha Touch, Phonegap/Apache
Cordova and Appcelerator Titanium.
After the writing of the code, following main components
of an app (views, controllers and data), an app could be
packaged for a native provision. The app packaging
process is similar for iOS or Android devices. Each
platforms differs by the configuration file. iOS requires
the registration to the Apple iOS provisioning profiles
portal in order to obtain a certificate, identity devices and
an App ID. Also Android requires a certificate to sign the
application by using the Android SDK Manager. For
Blackberry and Windows Phone App, Sencha touch in the
last version (2.3) introduces tools and an API for working
with Cordova and Phonegap in order that these
frameworks could package the app for the use on these
devices. In any case, developing with Sencha Touch
33
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
requires a set of actions that are not as simple as it seems.
Probably using the entire Sencha ecosystem allows an
easily coding by means of visual tools, but the other
frameworks are quite expensive. If we focus on the open
source version, technical skills are required.
requires technical skills and the knowledge of the target
mobile platforms and their development kits.
3.1.3 Appcelerator Titanium
Appcelerator company provides two main products: the
Appcelerator platform designed to enterprise that provides
a single open, cloud-based platform to deliver native
cross-platform apps and Titanium designed to developers
to create cloud-connected native apps.
Titanium is a development environment that includes an
open source SDK to support various devices and mobile
operating systems APIs, an integrated development
environment (Titanium Studio), an MVC framework
(Alloy) and cloud services to deliver apps.
Titanium products are provided as free, but it is required
the registration to the site. Titanium Studio (Fig. 10) is a
complete product that requires technical skills in order to
be used. The software is installed on the local machine,
but require the signing to the platform in order to work.
3.1.2 Phonegap vs. Cordova
Abobe Phonegap is built on top of Apache Cordova that is
a platform providing API for building native mobile
applications using the HTML5 framework. The difference
is on how their packaging tools are implemented.
Phonegap provides a remote building interface at Adobe
Phonegap Build that emulates an app for a single platform
in the cloud.
Phonegap requires the installation of the Nodejs
framework in order to install all the packages needed for
the app creation. Then it is required the presence on the
local system of the different SDKs.
As the Fig. 9 shows, the creation and the execution of an
app by means of the Phonegap command line tool,
requires the sign up to the Phonegap Build portal and thus
of an Adobe ID account.
Fig. 9 Steps required for Phonegap/Cordova installation
Fig. 10 Titanium Studio
Cordova packaging tools allow to build apps on a local
computer. In any case it is necessary to have a certificate
and AppID for each platform for the market in which a
developer what to distribute the app. Apache Cordova
requires the installation of Node.js (the JS platform used
to build network applications), that is used to install
locally the Cordova software. When installation is
completed, the Cordova tools allow to create the main
structure of the app inside a specific directory. When it is
necessary to add platforms that are needed to build the
project for the target mobile platform (e.g., iOS, android,
wp8, blackberry). The ability to run the commands
depends on whether the local machine support each SDK
(e.g., a Windows-based machine lacks of the iOS SDK).
The cordova command allows to build the app for the
target platform (i.e., using the build option), and to test
the app on an emulator or a device (i.e., using the emulate
option). Also in this case, the development of an app
3.2 A comparison between some strategies and tools
A comparison between native, web and hybrid apps based
on some criteria rated on a scale of 0 to 5 is shown in Fig.
11. Native apps, HTML5 apps, Phonegap apps and
Titanium apps are compared according the following
criteria: the features of the app, the user experience and
the performance, the monetization, the distribution, the
updates, the cost, the platforms fragmentation and
security.
34
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Some research (e.g., Canalys) reveals how Google Play is
the store that has achieved the highest number of
download (51% of total downloads carried out on the App
Store, Google Play, Windows Phone and BlackBerry
Store), but the App Store is the one that receives the most
revenue getting 74% of the total profits of the four
analyzed store. Even if the download of apps is
expanding, the gain does not increase in proportion, since
the percentage of free download is approximately 90% of
all installed apps. Native app has thus a low score
compared to other solutions whereas in the first case the
strategy is imposed by the store.
Different scores are related to the updates that solve bugs
or made apps’ improvements. The web app has the
highest score given that the update is automatic, a
Phonegap app has an average score considered that is
encoded with web technologies. Conversely a native app
and a Titanium app have a low score considering that the
updates imply the user having to re-download the app or
must wait since also updates are subjected to the store’s
approval process.
The costs are related to the number of platforms to be
achieved and the type of target device. The costs
associated with the web app are lower (high score)
considered the large number of developers with
knowledge of web technologies and thus their
competition. The other type of app have a cost that
depends on the knowledge of the tools and the crossplatform software. The creation of an app supported by
more than two platforms involves a high cost when
developing a native app. But also in the web-based
approach, there is a requirement for the app depending
the different versions of the mobile browsers installed in
the mobile devices.
The platform’s fragmentation means that different
versions of the mobile operating system may or may not
enable a functionality. The fragmentation has few
disadvantages in web-based app, app average in hybrid
app and many issues in native apps. Since there may be
several versions of the same operating system (e.g.
Android) active on different devices, development must
take into account the different hardware configurations
which every manufacturer adopts. A native app created
for the latest version of the operating system may not
work on systems that are equipped with the previous
versions.
Also a web app manifests a fragmentation problem that
concerns the mobile web browser each of which has
various versions and different engines that may or may
not implement the language-specific content. In this case,
the developers are obliged to test a web-based app on
different devices and browsers. A Phonegap app is able to
manage this fragmentation since it supports multiple
Fig. 11 Evaluation of methods of analysis for application development
features.
As far as apps features, native (score 5) and hybrid (score
4) apps offer the possibility to access to the various
resources of the device directly or via plugins or modules
provided by cross-platform tools. HTML5-based web app
instead (score 1), despite the rapid evolution of the
technology, have not yet reached the same level as the
other and do not allow to exploit the full capabilities of
the device on which they operate.
Focusing on the user experience and the performance,
since the performance evaluation is done according to the
graphics and the load time of an app, native apps have the
highest score (5), followed by hybrid app Titanium (4),
while the others methods have the minimum score. Note
that a poor performance lead to app failure, since user
wants an app that runs fast and with high performance.
Such features till now are possible only on native apps.
The user experience of the app Titanium is very close to
that of a native app, while a Phonegap app or a web app
may have performance issues related, for example, to the
time of loading a page, maybe because of the number of
files relating to external style sheets or scripts to load.
The gain is undoubtedly linked to the method of
distribution in the store and that is why hybrid native apps
and have maximum score than HTML5 -based app. The
app store distribution lead to a greater visibility with a big
chance of being purchased and create a revenue for
developers, although the competition is very high. The
web app is distributed via the website and therefore is less
visible since it is related to its availability through the
search engines.
35
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
[4] M. Pilgrim, HTML5: Up and running, O’Reilly, 2010
[5] S. Pastore, Mobile platforms and apps cross-platforms
development tools, Int. Journal of Engineering Research and
Application www. ijera.com. ISSN : 2248-9622, Vol. 3,
Issue 6, Nov-Dec 2013, pp.521-531.
[6] T. Pierce, Appreneur: secrets to success in the App store,
Apress, 2013
[7] T. Ferguson, App store revenue, downloads continue upward
trajectory – Canalys, Apr. 2013, at url: Mobilewordllive.com
[8] H. Koekkoek, Publication: how the most successful apps
monetize their user bade, Distimo.com publication, March.
2013.
[9] C. O’Connor, Evolution of the app business model,
Boston.com Aug, 2013
[10] L. Lazzarin, Il Mobile Internet italiano vola a 1,25 miliardi,
Content e app a 620 milion, wireless4innovation magazine,
Jun 2013.
software platforms at the same time. Conversely, the
extension of a product created with the Titanium
framework to all these operating systems would be
extremely expensive.
Finally, the security aspect presents a high score for a
native and a Titanium app and a low score for a Phonegap
app or a web-based app. The source code of a web-based
app is easily recoverable and there are several ways to
attack the security of a web app. On the other hand there
are techniques to secure a web app. HTML5 provides the
ability to cache data within the browser, and considering
that the app manage data on the device, these data should
be adequately protected and encrypted. The app can use
the native API of its operating systems to encrypt the
stored data, but this is not possible to implement such
technology with web languages.
As widely stated, the programmer is placed in front of a
series of choices. The final decisions about the tools to be
adopted are a compromise between the features and the
skills needed.
Serena Pastore Researcher/technologist from 2007 in ICT at the
Italian National Institute of Astrophysics (INAF) and University
teaching professor in some Italian University. Current research
interest are on Internet and web standards focusing on the mobile
environment, distributed network paradigms (grid and cloud), open
source products.
4. Conclusions
Developing a mobile app is not simple as it appears and
requires a carefully analysis on the mobile ecosystem and
some decisions about the strategies in the development.
The choice of a cross-platform tool that allows to
distribute an app on more than one mobile operation
system is a choice that requires technical knowledge and
skills comparable to those required in the development of
a native app. This is particularly true when we focus on
open source tools. Solutions like Sencha Touch or
Cordova could really help to develop various version of
the same app targeted to different mobile platforms, but
require time and cost. Moreover the mobile world is an
evolving and interesting market, but the revenue for an
app developer is not so sure. Even in a research
environment where we work, it is necessary to make an
analysis between the costs related to the development and
distribution of an app and the return in term of user
satisfaction and knowledge of our activities.
References
[1] F. Sheares, Power management in mobile devices, newness,
Elsevier, 2008.
[2] L. Columbus, 2013 Roundup of Samrtphone and Tablet
forecasts and market estimates, Forbes, Jan. 2013. At url:
http://www.forbes.com/sites/louiscolumbus/2013/01/17/201
3-roundup-of-mobility-forecasts-and-market-estimates/
[3] S. Pastore, The role of open web standards for website
development adhering to the one web vision. International
Journal of Engineering and Technology (IJET), Vol.2,
No.11. November 2012. ISSN 2049-3444.
36
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Data Mining Awareness and Readiness in Healthcare
Sector: a case of Tanzania
Salim Amour Diwani1, Anael Sam2
1
Information Communication Science and Engineering, Nelson Mandela African Institution of Science and Technology,
Arusha, P.O.BOX 447, Tanzania
[email protected]
2
Information Communication Science and Engineering,
Arusha, P.O.BOX 447, Tanzania
[email protected]
properly utilize the huge medical data collected through
different medical processes. Stored medical data
collection is an asset for healthcare organizations if
properly utilized. Healthcare industry uses data mining
techniques to fully utilize the benefits of the stored
medical datasets. This study investigates the readiness to
adopt, explore and implement healthcare data mining
techniques in Tanzania. The healthcare industry is
growing rapidly. The healthcare expenditures are also
increasing very high. Various healthcare organizations
worldwide such as world health organization (WHO) are
trying to provide quality healthcare treatments at cheaper
costs. The government sectors are able to benefits the use
of healthcare data mining technology in many ways.
Examples providing self healthcare treatments, using
scientific medical knowledge to provide healthcare
services to everyone, minimize time to wait for the
medical treatments, minimizing the delay time in
providing medical treatment, providing various healthcare
treatments based on the patients needs, symptoms and
preferences.
Abstract
Globally the application of data mining in healthcare is great,
because the healthcare sector is rich with tremendous amount of
data and data mining becoming very essential. Healthcare sector
collect huge amount of data in daily basis. Transferring data into
secure electronic systems of medical health will saves lives and
reduce the cost of healthcare services as well as early discovery of
contiguous diseases with advanced collection of data. This study
explore the awareness and readiness to implement data mining
technology within healthcare in Tanzania public sector . This study
is triangulated adopted online survey using Google doc and offline
survey using presentation techniques through different hospital
and distributed the questionnaires to the healthcare professionals.
The issues explored in the questionnaire included the awareness of
data mining technology, the level of understanding of, perception
of and readiness to implement data mining technology within
healthcare public sector. In this study we will analyze the data
using SPSS statistical tool.
Keywords: Data Mining , KDD, SPSS, healthcare,
ICT,WHO, Healthcare, Dynamic Morphology, NM-AIST,
NHIF, Paradigms, Positivists, HIS
2. Problem Description and Objectives
In Tanzania there has been no literature that has
discussed the implementation of data mining technology
within healthcare in the public sector . Various studies
have shown that new technology have become popular
and have been accepted by many people worldwide. This
study seeks to address this problem by investigating the
knowledge about and readiness with regard to the
implementation of the data mining technology in
healthcare in the public sector. Prior studies of data
mining readiness and implementation have been
undertaken in the private sector. Evidence suggests that
personnel within private sector firms are aware and ready
1. Introduction
In recent years, there has been a tremendous change in
the adaptation and implementation of new technologies in
both private and public sectors. Data mining is an
emerging technology used in different types of
organizations to improve the efficiency and effectiveness
of business processes. Data mining techniques analyze
large data sets to discover new relationships between the
stored data values. Healthcare is an information rich
industry, warehousing large amount of medical data.
Health care industry finds it difficult to manage and
37
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
to implement this technology. Studies in the area of
telecommunication, banking, and insurance companies
indicate that there is a level of optimism and
innovativeness among employees indicating the potential
to adopt data mining techniques. Readiness can be seen in
terms of the adoption of or intent to adopt data mining
technologies [1],[2],[3],[4].
disease and symptoms, curbing the treatment cost,
predicting medical diagnosis and can help physicians to
discover which medications are most effectively and
efficient for patients. Healthcare Data mining uses
dynamic morphology for mining healthcare data.
Dynamic morphology is the process of extracting
healthcare data in order to discover new knowledge and
relationships. Refer to figure 2 below, let say we want to
mine the tuberculosis data using dynamic morphology we
need to follow the following procedures:
 analyze stored medical data for TB
3. Data Mining
Data Mining is the process of extracting useful
information from large and mostly unorganized data
banks. It is the process of performing automated
extraction and generating predictive information from
large data banks. It enables you to understand the current
market trends and enables you to take a proactive
measures to gain maximum benefit from the same. Refer
to figure 1 below, data mining is a step in the knowledge
discovery in databases (KDD) process and refers to
algorithms that are applied to extract patterns from the
data. The extracted information can then be used to form
a prediction or classification model, identify trends and
associations, refine an existing model, or provide a
summary of the database being mined[5].

identify causes or symptoms of TB

Transform discovered knowledge using any one
of the healthcare data transformation techniques
to explore extra information

store the results of knowledge transformation for
further analysis
Fig. 2 Example of Dynamic Morphology for Tuberculosis
Fig. 1 KDD Process
4. Data Mining in Healthcare
5. Data Mining Readiness
The healthcare industry is growing abundantly. The
healthcare expenditures are also increasing tremendously.
Healthcare organization worldwide such as National
Health Insurance Fund (NHIF) in Tanzania are trying to
provide quality healthcare at cheaper cost. The healthcare
organizations are adopting new technologies, which will
help in early detection of life threatening diseases and
lowering the medical cost. Data mining in healthcare is
useful in segmenting patients into groups, identifying
patients with recurring health problems, relation between
In order to accept and ready to apply or adopt data mining
techniques, the major issues is the willingness and
capability to accept the technology. Human resources
primarily on their readiness toward accepting data mining
technology can be argued as a major obstacle in adopting
a new technology within any organization[3]. In this
study issues whether people use such a technology or not
and their ready to accept a technology or not are major
concern. Hence, in order to discover the readiness towards
data mining technology issue like clarity for business
38
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
strategy, user skills and experiences, organization and
culture, technology and data quality needs to be
considered.
6.2 Study Settings and Time Horizon
The setting selected for the study were hospitals in Dares-salaam, Tanga, Zanzibar, Arusha and Dodoma. This
study was undertaken via presentation through the use of
survey in the local hospital during their breakfast
meeting. The questionnaire was distributed among
respondent and later the respondent were contacted. The
process of distributing and collecting the questionnaire
begins in January 2013 and end in March 2013.
6. Research Design and Methodology
7. Purpose and Approach
The study aims to gain an understanding of data mining
technology utilization in Tanzania healthcare public
sector. The impact factors which data mining technology
can be sought to be utilized such as level of awareness and
readiness of data mining technology in healthcare public
sector. Since there is little or no any paper which has been
published in data mining technology in Tanzania. Also
there are very few companies who implement and use data
mining technology in their day to day activities such as in
decision making process. This study aim to discover the
level of awareness and readiness in data mining
technology in healthcare sector and the willingness to
accept or reject the implementation. This study will assist
in filling the gap in the literature as well.
Figure 3: Modeling the research design [6]
The above figure explore the study design which will be
adopted in this study. The design consists of research
questions which need to be answered and analyzed,
identify participants, location of the study and time
horizon, purpose and justification of the study before
discussing the research paradigm and approach
considered. Each of this part of the research design will
be discussed in detailed.
7.1 Paradigm
All researchers have different beliefs and ways of looking
and interacting within their environment. As a result, the
way in which research is conducted vary. However, there
are certain standards rules that govern researchers
actions and beliefs. A paradigm is “a broad view or
perspective of something”[7]. Hence, definition of
paradigm reveals how research could be affected and
guided by a certain paradigm by stating, “paradigms are
patterns of beliefs and practices that regulate inquiry
within a discipline by providing lenses, frames and
processes
through
which
investigation
is
accomplished”[8]. This study utilized a positivist
approach which explore social reality, such as attitudes of
readiness toward data mining and their satisfaction
toward healthcare system. The concept is based on the
philosophical ideas of emphasizing observation and
reason behind participant actions in adopting data mining
technology for example, their knowledge and
understanding about data mining, how they think about
technology advancement in their work place.
6.1 Participants
Participants in this study are professionals who working
in medical fields in Tanzania public sectors. The
respondent selected are physician, medical doctors,
dentists, nurses, pharmacists, and hospitals body of
directors. In this study individuals who are working in
medical sectors are surveyed to address the basic research
problem in the level of awareness and readiness to adopt
data mining techniques. The study also seek if people are
ready or not to adopt data mining techniques in
healthcare. The questionnaires were prepared and
available in English through Google docs and posted
through emails of individuals who are working in medical
field. Also the questionnaires were distributed manually
in the hospital and research centers after conducting
presentation which explained more about the study
design.
39
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
approximate intake of 500 students, also we have more
than 6000 medical doctors scattered around the country[8]
. Structured questionnaires were designed and online
survey was conducted to reach the targeted sample size of
400 respondents out of the entire population of the at least
6000 medical specialists. Different medical specialists
(surgeon,
gynecologists,
physicians),
doctors,
Pharmacists, Nurse and information system officers were
also involved in the survey.
7.2 Approach
This research is to be conducted in African developing
countries. Consequently I expect issues such as national
culture, form of government, and the politics of protest
and shape the nature of this study in a different way as
compared to a study conducted in the context of developed
countries. Hence, this study utilized a triangulation
approach to explore the survey data which will provide
basic understanding about what happening in the industry
in regard to data mining technology such as data mining
awareness, attitudes towards data mining technology and
software used related to data mining if any within
healthcare sector. Also, if healthcare professionals are
willing to adopt data mining technology and able to
upgrade their skills.
8. Analysis and Findings
Refer to figure 4, 5 and 6 below, awareness and readiness
of the respondents were assessed if they know or do not
know any other health information systems and other
questions were specific to data mining tool. 60% (241) of
the respondents knew at least one health information
system excluding data mining tool, 40% (158) did not
know any health information system. Among them, 53%
(211) respondents have once used it meanwhile 47%
(189) have never used any health information system.
In the organization’s perspective, 48% (190) of the
organizations where respondents once worked at have an
experience in a health information systems excluding
data mining tool while 50% (202) of the organizations
had no experience in health information systems and 2%
(8) used other systems.
In order to gain more understanding of the respondents’
readiness in adopting technology particularly in data
mining, the following statements were provided to the
respondents, these include whether technology gives the
respondents greater control over his/her daily work
activities, to what extent do they agree that technologies
are much more convenient in the products and services,
preference in using the most advanced technology
available, efficient of technology in the respond’s
occupation, to keep up with the latest technological
developments in his/her areas of interest, to have fewer
problems than other people in making technology
working for them, flexibility to learn about new and
different technology, usefulness of the technology for
accomplishment of any tasks and willingness to use data
mining technology for analyzing medical data in addition
to current methods. All these statements had to be
responded in the rate of scale which includes strongly
disagree, agree, neutral, Disagree and strongly disagree.
Strongly disagree and disagree were put by a researcher
to be negative response category and agree and strongly
agree were put in the category of positive category. All
the results are presented in the Chart below:
7.3 Instrument Design
Questionnaires was designed as a quantitative part of this
study using Google doc. There is no wrong or right
answers in the questionnaire what we try to test is the
level of awareness regard to data mining technology in
healthcare and if the healthcare practitioners are willing
to adopt a new technology or not. The design of the
questionnaires of this study adopted several sources of
data, including previous instrument developed by other
researchers and literature related to data mining adoption.
Types of questions were closed ended questions with
ordered and unordered responses using Likert type scale
like level of measurement for example 1=strongly
disagree and 5 = strongly agree. Itemized rating scale was
also developed for few questions. For example, scale from
1=poor to 5=excellent was used for question as to rate the
actual performance on important factors for quality
healthcare system. There is also similar scale 1=seldom to
5=very often was used for question requesting to indicate
frequency of healthcare data were used in particular
areas. The questionnaires were pre-tested by sending
through the emails of NM-AIST colleagues to review the
questions critically and getting the feedback from them
about the content and form, and how to make it more
appropriate and were to send it, in order to get valuable
feedback.
7.4 Data Collection
A survey was conducted for 400 respondents from
different parts of the country including Dar es salaam,
Arusha, Zanzibar, Tanga and Dodoma. In Tanzania we
have 5 medical University in operation with
an
40
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
users will catch up with it slowly because it will not be
easy for them to adapt to it quickly. Also, the majority of
the respondents stayed neutral about new technology like
data mining. This is due to limited knowledge of data
mining tool.
Figure 6: Neither positive nor negative toward the factor
Figure 4: Positive response in percentage supporting the factor
9. Reliability
In this section the reliability of data in the study is
discussed. Refer to table 1, 2 and 3 below, a reliability test
was conducted using Cronbach’s Alpha to measure the
internal consistence or average correlation of the items. .
It ranges from 0-1. The most reliable value is 0.7. In this
study alpha obtained after analysis in SPSS was 0.974
which means the correlation of the research questions
were most appropriate. The key variables used in the
statistical analysis are: data mining readiness and
awareness.
Table 1: Health Information System
Scale Mean
Scale
Corrected
Cronbach's
if Item
Variance if
Item Total
Alpha if
Deleted
Item Deleted
Correlation
Item
Deleted
Do you
1495.5556
2614954.365
.304
.974
1495.5000
2614827.802
.307
.974
know any
HIS in
Tanzania?
Figure 5: Negative response in percentage not supporting the factor
Have you
It is observed that new technology like data mining tools
is essential to facilitate personal performance though
ever used
any HIS?
41
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Table 2: Data Accuracy
Table 3: Medical Data
Scale Mean
Scale
Corrected
Cronbach's
Scale
if Item
Variance if
Item-Total
Alpha if
Mean
Deleted
Item Deleted
Correlation
Item
Item
Total
Item
Deleted
Deleted
Correlati
Deleted
The data
conforms to
Correcte
Cronbach's
if Item Deleted
d
Alpha
Item-
if
on
Please
1489.5000
2594194.972
indicate how
.473
frequently you
the actual
use
value
medical data
data which
from HIS in
recorded in
each of the
your system
Scale Variance
.974
which
recorded is
if
1489.4074
2594545.454
.469
the
1465.6667
2548146.038
.542
.974
1462.0370
2536410.263
.608
.973
following
.974
areas-
is up to date
Decision
Complete :
Making
all
data analysis
relevance
Please
value for a
1487.9815
2597785.792
.366
and
indicate how
.974
certain
frequently you
variable are
use
recorded
medical data
Consistent:
from HIS in
representati
each of the
following
on of the
data value
the
1489.8519
2593776.166
.477
areas-
.974
is the same
Performance
in all
measurement
classes
and planning
Please rate
the actual
performanc
10. Discussion
e on each of
those
1438.9630
2568839.206
.374
.974
Data mining in healthcare helps to improve healthcare
standards. Data mining technology helps healthcare
organizations to achieve various business goals such as
low costs, increase revenue generation while maintaining
high quality healthcare. Data mining in healthcare
identified as the process of extracting and analyzing
healthcare data for presentation in a format allowing the
generation of information and the creation of knowledge
through the analysis of information in order to enhance
the decision making process within the public sector
organization.
The major research problem addressed in this study was
in Tanzania there is a lack of the status of implementation
factors by
your
organizatio
n
42
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Mandela African Institute of Science and Technology and
Dr. Muhammad Abulaish from Jamia Milia Islamia for
their guidance and valuable support during this work.
of data mining techniques within healthcare in the public
sector and benefits to be derived by implementing such
technologies. The study sought to increase an
understanding of data mining technology in Tanzania
within public sector. Areas of interests within this study
were data mining awareness and readiness to accept or
reject technology in healthcare public sector, the reasons
to accept or reject the technology and impacts of data
mining technologies.
References
[1] Berger, C. (1999). ‘Data mining to reduce churn’.
Target Marketing, 22(8), 26.
[2] Chye, K. H., & Gerry, C. K. L. (2002). ‘Data Mining
and Customer Relationship Marketing
in the Banking
Industry’. Singapore Management Review, 24(2), 1-27.
[3] Dahlan, N., Ramayah, T., & Hoe, K. A. (2002). Data
Mining in the Banking Industry: An Exploratory Study.
The proceedings of the International Conference 2002.
Paper presented at the ‘Internet Economy and Business’,
Kuala Lumpur, 17-18th September 2002.
[4] Chun, S.-H., & Kim, S. H. (2004). ‘Data mining for
financial prediction and trading: application to single and
multiple markets’. Expert System with Applications, 26,
131-139.
[5] Fayyad, U., Piatetsky-Shapiro, G., & Smyth, P.
(1996). ‘The KDD process for extracting useful
knowledge from volumes of data’. Association for
Computing Machinery. Communications of the ACM,
39(11), 27-34.
[6] Cavana, R. Y., Delahaye, B. L., & Sekaran, U. (2001).
Applied Business Reseach:Qualitative and Quantitative
Methods. Sydney and Melbourne: John Wiley & Sons
Australia, Ltd.
[7] Taylor, B., Kermode, S., & Roberts, K.(2007).
Research in nursing and health care: Evidence for
practice. (3rd ed.). South Melbourne, VIC: Thomson.
[8] Weaver, K., & Olson, J.K.(2006). Understanding
paradigms used for nursing research. Journal of Advanced
Nursing, 53(4), 459-469.
11. Conclusion
This paper discussed the awareness and readiness to adopt
data mining techniques in healthcare sector in Tanzania.
The statistical results of the questionnaire were collected
and analyzed, and the issues discussed in the
questionnaire were healthcare information system, data
mining
readiness,
data
mining
technologies
implementation and perception of data mining. The
descriptive statistics has shown that many healthcare
professionals in Tanzania they are not aware of the data
mining technology and they never used it before. Hence
they are willing to accept and adopting data mining
technology if they get an opportunity to do so.
Respondents identified the influencing factors for
utilizing data mining and reason for not utilizing it. Each
of the major research questions were analyzed and the
results were given. Cronbach’s Alpha was used to test the
internal consistence of the questions; the alpha test was at
least 0.7 which means the questions were most reliable.
All the questions used likert scales, those scales which
were coded as strongly agreed and agreed were considered
to measure positive response while disagreed and strongly
disagreed were considered to measure negative response.
The findings showed the majority are aware of health
information system but limited knowledgeable on data
mining technology. About performance of health
information system the majority of the respondents
strongly agreed that the system should be easy to use, able
to validate data, adequate and sufficient documentation
for employees to follow and easy to modify and upgrade
information. Also they suggested that sufficient financial
resources, effective and adequate for the staffs is required
in order to implement data mining in developing
countries like Tanzania.
Salim Amour Diwani received his BS degree in computer
science at Jamia Hamdard University, New Delhi, India in
2006 and Ms degree in computer science at Jamia
Hamdard University in New Delhi, India in 2008. He is
currently a PhD scholar in Information communication
Science and Engineering at Nelson Mandela African
Institution of Science and Technology in Arusha, Tanzania.
His primary research interests are in Data Mining, Machine
Learning and Database Management Systems.
Acknowledgement
I would like to thank almighty and my family for their
constant support during this work. I would also like to
thank my supervisors Dr. Anael Sam from Nelson
43
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Digital Organism Simulation Environment (DOSE): A Library
for Ecologically-Based In Silico Experimental Evolution
Clarence FG Castillo1, and Maurice HT Ling 2
1
2
School of Information Technology, Republic Polytechnic
Singapore, Republic of Singapore
[email protected]
School of Chemical and Biomedical Engineering, Nanyang Technological University
Singapore, Republic of Singapore
Department of Zoology, The University of Melbourne
Parkville, Victoria 3010, Australia
[email protected]
Richard Lenski in 1988 [4], using a common intestinal
bacterium, Escherichia coli, which has one of the shortest
generation time. Other experimental evolution
experiments [5-7], such as adaptation to salt and food
additives, have also used E. coli due to its generation time.
Despite so, it is generally prohibitively expensive to
examine the genetic makeup of each bacterium using
experimental techniques, such as DNA sequencing. At the
same time, such examination is destructive in nature and
the examined bacterium cannot be revived for further
evolutionary experiments.
Abstract
Testing evolutionary hypothesis in biological setting is
expensive and time consuming. Computer simulations of
organisms (digital organisms) are commonly used proxies to
study evolutionary processes. A number of digital organism
simulators have been developed but are deficient in biological
and ecological parallels. In this study, we present DOSE
(Digital Organism Simulation Environment), a digital organism
simulator with biological and ecological parallels. DOSE
consists of a biological hierarchy of genetic sequences, organism,
population, and ecosystem. A 3-character instruction set that
does not take any operand is used as genetic code for digital
organism, which the 3-nucleotide codon structure in naturally
occurring DNA. The evolutionary driver is simulated by a
genetic algorithm. We demonstrate the utility in examining the
effects of migration on heterozygosity, also known as local
genetic distance.
Keywords: Digital Organisms, Simulation Environment,
Ecology Simulation, Migration, Genetic Distance.
A means around these limitations is to use models of
bacteria or higher organisms, rather than real biological
organisms. These modeled organisms are known as
artificial life or digital organisms (DO) which organisms
are simulated, mutated, and reproduced in a computer [8].
Although digital organisms are not real biological
organism, it has characteristics of being a real living
organism but in a different substrate [9]. Batut et al. [3]
argue that DO is a valuable tool to enable experimental
evolution despite its drawbacks as repeated simulations
can be carried out with recording of all events.
Furthermore, only computational time is needed to study
every organism, which is analogous to sequencing every
organism, and this process is not destructive in a
biological sense as the studied organism can be “revived”
for further simulations.
1. Introduction
Nothing in Biology makes sense except in the light of
Evolution -- Theodosius Dobzhansky [1]
Nothing in Medicine makes sense, except in the light of
Evolution -- Ajit Varki [2]
Evolution is a fundamental aspect of biology. However,
testing evolutionary hypotheses is a challenge [3] as it is
highly time consuming and expensive, if not impossible.
Long generation time associated with most species makes
it virtually impossible to test evolutionary hypotheses in a
laboratory setting. The longest on-going laboratory
experiment in evolutionary biology have been initiated by
The main tool needed for using DO is a computational
platform to act as a simulation environment. A number of
DO platforms have been developed [10]. One of the early
simulators is Tierra [11], where each organism is an
evolvable, mating and reproducing program competing
for computing resources, such as CPU cycles and memory
44
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
The 4 sets of components are briefly described as follow;
firstly, DOSE consists of a set of objects representing a
chromosome, organism, and population [17]. An
organism can consist of one or more chromosome to make
up its genome and a population consists of one or more
organisms. Secondly, a GA acts as the evolutionary driver
acting on the chromosomes. Thirdly, Ragaraja interpreter
[19] is used to read the chromosomes and update the
cytoplasm (cell body). This resembles the translation of
genes into proteins in biological context; hence, Ragaraja
interpreter [19] can be seen as the transcription and
translation machinery. Lastly, a 3-dimensional world [18]
consisting of ecological cells allows the mapping of DOs
onto the world.
space. Hence, Tierra’s programs can be seen as an
executable DNA. A major drawback of Tierra is that the
DOs are not isolated from each other as all DOs shared
and compete for the same memory space. Avida [12]
simplified Tierra [11] by enabling each DO to run on its
own virtual machine; thus, isolating each DO, resulting in
CPU cycle being the main competing resource. As Tierra
[11] and Avida [12] used bytecodes as the genetic
constituents for DO, it is difficult to examine parameters
such as heterozygosity and genetic distance, which is
commonly used in population genetics [13] from HIV
virus [14] to human migration [15]. Mukherjee et al. [16]
defines heterozygosity as variation within population
while genetic distance is the variation between
populations. Hence, heterozygosity can be considered as
local genetic distance or within group genetic distance.
Aevol [3] used a binary string as genetic material and had
incorporated concepts from molecular biology; such as
genes, promoters, terminators, and various mutations;
into its design. This allowed for genetic distance to be
measured. However, aevol [3] is designed for simulating
bacterial genetics and evolution. Hence, ecological
concepts, such as migration and isolation, are not
incorporated.
Each simulation is defined by a set of parameters and
functions, which are used by the simulation driver and
management. It constructs and initializes the DOs, maps
the DOs onto the world, runs the simulation from first
generation to the maximum generation as defined in the
parameter, and report the events into a text file or
database as required. After DO initialization, the current
simulation driver simulates each organism and ecological
cell sequentially [18].
The following is the core set of 18 parameters available in
DOSE to cater for various uses:
 population_names: provides the names of one or more
populations
 population_locations: defines the deployment of
population(s) at the start of the simulation
 deployment_code: defines the type of deployment
scheme
 chromosome_bases: defines allowable bases for the
genetic material
 background_mutation: defines background mutation
rate
 additional_mutation: defines mutation rate on top of
background mutation rate
 mutation_type: defines a default type of mutation
 chromosome_size: defines the initial size of each
chromosome
 genome_size: defines the number of chromosome(s) in
each organism
 max_tape_length: defines the size of cytoplasm
 interpret_chromosome: defines whether phenotype is to
be simulated
 max_codon: defines the maximum number of codons to
express
 population_size: defines the number of organisms per
population
Previously, our group had designed a genetic algorithm
(GA) framework conforming to biological hierarchy
starting from gene to chromosome to genome (as
organism) to population [17], which may help
interpreting GA results to biological context. Further
work [18, 19] by our group had formalized a 3-character
genetic language to correspond the 3-nucleotide codon in
naturally occurring DNA and incorporating a 3dimensional “world” consisting of ecological cells in
order to give it parallels to biological DNA and natural
ecosystem.
Here, we present a Python DO simulation library, Digital
Organism Simulation Environment (DOSE), built on our
previous work [17-19]. We then illustrate the use of
DOSE to examine the effects of migration on
heterozygosity (local genetic distance) where DOs can
only mate within their own ecological cell.
2. Methods
2.1 DOSE Library
The basis of DOSE is a simulation driver and
management layer built on top of 4 different sets of
components, which had been previously described [17-19].
45
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
were not allowed to cross cell boundaries throughout the
simulation in order to simulate complete isolation. In
adjacent migration simulation, 10% of the organisms from
a cell can migrate to one of its 8-neighbour cell within a
generation in order to simulate short distance migration
patterns, such as foraging or nomadic behavior. In long
migration, 10% of the organisms from a cell can migrate to
any other cells within a generation in order to simulate long
distance migration patterns, such as flight. Each simulation
was performed for 1000 generations.
 world_x, world_y, world_z: defines the size of the world
in terms of numbers of ecological cells
 maximum_generations: defines the number of
generations to simulate
 ragaraja_instructions: list of recognized codons
The following is the core set of 12 functions definable in
DOSE to cater for various uses; of which, Functions 2 to
11 were previously defined [18]:
1. deployment_scheme: initial deployment of organisms
into the ecosystem
2. fitness: calculates the fitness of the organism and
returns a fitness score
3. mutation_scheme: mutation events in each
chromosome
4. prepopulation_control: population control events
before mating event in each generation
5. mating: mate choice and mating events
6. postpopulation_control: population control events
after mating event in each generation
7. generation_events: other irregular events in each
generation
8. organism_movement: short distance movement of
organisms within the world, such as foraging
9. organism_location: long distance movement of
organisms within the world, such as flight
10. ecoregulate: events to the entire ecosystem
11. update_ecology: local environment affecting entire
ecosystem
12. update_local: ecosystem affecting the local
environment
2.3 Data Analysis
Within cell analyses were performed. Hamming distance
[20] was calculated between the chromosomes of two
organisms and used as local genetic distance. 50 random
pairs of organisms within a cell were selected for pairwise local genetic distance calculation and an average
heterozygosity was calculated for each cell in every
generation. Within a generation, mean and standard error
of heterozygosity were calculated from the average local
genetic distances of 25 cells for each simulation.
3. Results
In this study, we present Python DO simulation library,
Digital Organism Simulation Environment (DOSE), built
on our previous work [17-19]. We first briefly outline a
typical use of a DO simulation platform such as DOSE
before illustrating 2 examples to examine the effects of
migration on heterozygosity, given that the DOs can only
mate within their own ecological cell.
2.2 Simulations
3.1 Typical use of an in silico evolutionary platform
Two sets (Example 1 and Example 2) of three simulations
with different migration schemes; no migration, adjacent
migration, and long migration; were developed, giving a
total of six simulations. Each simulation consisted of a 25cell flat world with 50 organisms per cell and mating could
only be carried out between organisms within the same cell.
As a result, each cell resembled an isolated landmass. One
binary chromosome of 5000 bases formed the genetic
material for each organism. Only point mutation was used
and the two sets of simulation differ by point mutation rates.
In the first set of 3 simulations (Example 1), mutation rate
was set at 0.001, resulting in 5 point mutations per
generation. In the second set of simulations (Example 2),
mutation rate was set at 0.002, effectively doubling the
occurrence of point mutations per generation compared to
Example 1. Since the chromosomes were binary, mutation
events were limited to inverting the base from one to zero
and vice versa. Mutation scheme was identical in all 3
migration schemes. In no migration simulation, organisms
Similar to other in silico evolutionary platforms such as
aevol [3], the basic output of DOSE is a set of time series
data with generation count as the timing variable. These
can include organism statistics; such as fitness, and
genome size; or population statistics; such as average
fitness, and genetic distance. Further analyses can be
carried out from these results. For example, if the parentchild (also known as ancestry) relationships are recorded,
the lineage of beneficial mutations can be carried out
using genealogical analysis [21]. Further studies using 2
or more evolved populations of digital organisms, such as
measuring mutational robustness using a competition [22,
23], may be performed. These competition assays may be
used to model biological processes, such as parasitism
[24].
A typical in silico evolutionary experiment consists of
modifying one or more parameters, such as mutation rate,
46
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
migration between sub-groups of mating populations
results in the increase of genetic variation within local
populations. The scenario of no migration acts as a
control and long migration scenario yields the same local
population variation as control where genetic variation
only occurs from mutations. This suggests that long
distance migration covering the entire ecosystem may
result in the entire ecosystem behaving as one
geographically extensive “local” population. This is
observed in hoverflies where extensive migration result in
the lack of genetic differentiation in a continental scale
[29]. A study in human populations also suggested that
long migration may result in the lack of genetic variation
between sub-populations [30], which is consistent with
our simulation results.
and/or functions, such as mating scheme, in the platform,
and examining the time series data emerging from one or
more simulations. Batut et al. [3] highlighted that
fortuitous events can be distinguished from systematic
trends by comparing data from replicated simulations. It
is also possible to revive one or more simulations from
stored data and that can be mixed to simulate interactions
between groups of organisms [25].
3.2 Example 1: Testing the effects of migration on
heterozygosity
DOSE is designed as a tool to examine evolutionary
scenarios on an ecological setting. In this example, we
examine the effects of migration, simulated by movement
of organisms to adjacent or across non-adjacent ecological
cells.
Hamming distance [20], which had been used as distance
measure for phylogenetic determination between viruses
[26, 27], was used in this study as a measure of
heterozygosity. As chromosomal lengths were identical in
all organisms throughout the simulation, Hamming
distance represented the number of base differences
between any two organisms.
Our results show that the average heterozygosity for no
migration and long migration across all 1000 generations
for all 25 ecological cells is similar (p-value = 0.989;
Table 1). The average heterozygosity for adjacent
migration is marginally lower but not significantly
different from that of no migration (p-value = 0.932) or
long migration (p-value = 0.921). The average spread
(standard error) of heterozygosity for no migration and
long migration is also similar (p-value = 0.264; Figure 1A
and 1C). However, the spread of heterozygosity for
adjacent migration is significantly larger (p-value < 4.3 x
10-26), especially after 500 generations (Figure 1B).
Fig. 1a Standard error of heterozygosity for no migration scenario.
Table 1: Summary statistics of 3 simulations with mutation rate of 0.001
Simulation
Average
Heterozygosity
Average
Standard Error
No migration
1228.19
25.064
Adjacent migration
1226.09
32.664
Long migration
1228.54
25.661
The average spread of heterozygosity from organisms
within an ecological cell can be used as a proxy to
estimate the variation within local population or intrapopulation [28]. Our results suggest that adjacent
Fig. 1b Standard error of heterozygosity for adjacent migration scenario.
47
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Fig. 1c Standard error of heterozygosity for long migration scenario.
Fig. 2 Standard errors of heterozygosity between no migration and long
distance migration for mutation rate of 0.002.
Our results also suggest that migration and mating
between adjacent sub-populations increased the genetic
variability, as seen in increased variation between
adjacent migration and no migration scenarios. This is
supported by current study suggesting that migration is
crucial in maintaining genetic variation [31].
By comparing simulation outputs from different mutation
rates (0.1% against 0.2%), our results show that
heterozygosity (Figure 3A) and spread of heterozygosity
(Figure 3B) are increased with higher mutation rate. This
increase is significant for both heterozygosity (p-value <
6.8 x 10-90) and spread of heterozygosity (p-value < 7.3 x
10-55). However, the trend is consistent in both examples.
This is consistent with Mukherjee et al. [16] whom
demonstrates that mutation rates does not impact on the
statistical tests for evaluating heterozygosity and genetic
distance using a simulation study.
3.3 Example 2: Testing the effects of mutation rates
and migration on heterozygosity
In this example, we double the mutation rate from 0.001
(0.1%) to 0.002 (0.2%) on the 3 migration scenarios in
Example 1. The simulation results can be analyzed in the
same manner as Example 1 or compared with that of
Example 1 to examine the effect of increased mutation
rate.
Our results show that there is no difference in the average
heterozygosity between all 3 simulations (F = 0.01, pvalue = 0.987; Table 2). The spread of heterozygosity is
significantly higher in adjacent migration when compared
to no migration (p-value = 4.4 x 10-34) or long migration
(p-value = 2.2 x 10-31) scenarios (Figure 2). These results
are consistent with that of Example 1, suggesting that
these trends are not significantly impacted by doubling
the mutation rate.
Table 2: Summary statistics of 3 simulations with mutation rate of 0.002
Simulation
Average
Heterozygosity
Average
Standard Error
No migration
1787.79
36.296
Adjacent migration
1784.32
45.776
Long migration
1788.52
36.695
Fig. 3a Mean heterozygosity between migration scenarios for both mutation
rates.
48
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Infocomm Research, Singapore) for their comments on an
initial draft of this manuscript.
References
[1] T. Dobzhansky, “Nothing in biology makes sense except in
the light of evolution”, The American Biology Teacher, Vol.
35, 1973, pp. 125-129.
[2] A. Varki, “Nothing in medicine makes sense, except in the
light of evolution”, Journal of Molecular Medicine, Vol. 90,
2012, pp. 481-494.
[3] B. Batut, D. P. Parsons, S. Fischer, G. Beslon and C. Knibbe,
“In silico experimental evolution: a tool to test evolutionary
scenarios”, BMC Bioinformatics, Vol. 14, No. Suppl 15,
2013, Article S11.
[4] J. E. Barrick and R. E. Lenski, “Genome dynamics during
experimental evolution”, Nature Reviews Genetics, Vol. 14,
No. 12, 2013, pp. 827-839.
[5] C. H. Lee, J. S. H. Oon, K. C. Lee and M. H. T. Ling,
“Escherichia coli ATCC 8739 adapts to the presence of
sodium chloride, monosodium glutamate, and benzoic acid
after extended culture”, ISRN Microbiology, Vol. 2012,
2012, Article ID 965356.
[6] J. A. How, J. Z. R. Lim, D. J. W. Goh, W. C. Ng, J. S. H.
Oon, K. C. Lee, C. H. Lee and M. H. T. Ling, “Adaptation
of Escherichia coli ATCC 8739 to 11% NaCl”, Dataset
Papers in Biology 2013, 2013, Article ID 219095.
[7] D. J. W. Goh, J. A. How, J. Z. R. Lim, W. C. Ng, J. S. H.
Oon, K. C. Lee, C. H. Lee and M. H. T. Ling, “Gradual and
step-wise halophilization enables Escherichia coli ATCC
8739 to adapt to 11% NaCl”, Electronic Physician, Vol. 4,
No. 3, 2012, pp. 527-535.
[8] S. F. Elena and R. Sanjuán, “The effect of genetic robustness
on evolvability in digital organisms”, BMC Evolutionary
Biology, Vol. 8, 2008, pp. 284.
[9] Y. Z. Koh, and M. H. T. Ling, MHT, “On the liveliness of
artificial life”, Human-Level Intelligence, Vol. 3, 2013,
Article 1.
[10] A. Adamatzk, and M. Komosinski M, Artificial life models
in software, London: Springer-Verlag, 2005.
[11] T. S. Ray, 1991, "Evolution and optimization of digital
organisms", in Billingsley K.R. et al. (eds), Scientific
Excellence in Supercomputing: The IBM 1990 Contest Prize
Papers, 1991, pp. 489–531.
[12] C. Ofria, and C. O. Wilke, “Avida: A software platform for
research in computational evolutionary biology”, Artificial
Life, Vol. 10, 2004, pp. 191-229.
[13] O. Tal, “Two complementary perspectives on interindividual genetic distance”, Biosystems, Vol. 111, No. 1,
2013, pp. 18-36.
[14] M. Arenas, and D. Posada, “Computational design of
centralized HIV-1 genes”, Current HIV Research, Vol. 8,
No. 8, 2010, pp. 613-621.
[15] S. J. Park, J. O. Yang, S. C. Kim, and J. Kwon, “Inference
of kinship coefficients from Korean SNP genotyping data”,
BMB Reports, Vol. 46, No. 6, 2013, pp. 305-309.
[16] M. Mukherjee, D. O. Skibinski, and R. D. Ward, “A
simulation study of the neutral evolution of heterozygosity
Fig. 3b Standard error of heterozygosity between migration scenarios for
both mutation rates.
4. Conclusions
In this study, we have presented a Python DO simulation
library, Digital Organism Simulation Environment
(DOSE), built on our previous work [17-19]. DOSE is
designed with biological and ecological parallels in mind.
As a result, it is relatively easy to construct evolutionary
simulations to examine evolutionary scenarios, especially
when a complex interaction of environment and biology is
required. To illustrate the use of DOSE in an ecological
context, we have presented 2 examples on the effects of
migration schemes on heterozygosity. Our simulation
results show that adjacent migration, such as foraging or
nomadic behavior, increases heterozygosity while long
distance migration, such as flight covering the entire
ecosystem, does not increase heterozygosity. These results
are consistent with previous studies [29, 30].
Appendix
DOSE version 1.0.0 is released under GNU General
Public
License
version
3
at
http://github.com/mauriceling/dose/
release/tag/v1.0.0
and anyone is encouraged to fork from this repository.
Documentation
can
be
found
at
http://
maurice.vodien.com/project-dose.
Acknowledgement
The authors will like to thank AJZ Tan (Biochemistry,
The University of Melbourne) and MQA Li (Institute of
49
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
[31] M. Yamamichi, and H. Innan, “Estimating the migration
rate from genetic variation data”, Heredity, Vol. 108, No.
4, 2012, pp. 362-363.
and genetic distance”, Heredity, Vol. 53, No. 3, 1987, pp.
413-423.
[17] J. Z. R. Lim, Z. Q. Aw, D. J. W. Goh, J. A. How, S. X. Z.
Low, B. Z. L. Loo, and M. H. T Ling, “A genetic algorithm
framework grounded in biology”, The Python Papers Source
Codes, Vol. 2, 2010, Article 6.
[18] M. H. T Ling, “An artificial life simulation library based on
genetic algorithm, 3-character genetic code and biological
hierarchy”, The Python Papers, Vol. 7, 2012, Article 5.
[19] M. H. T Ling, “Ragaraja 1.0: The genome interpreter of
Digital Organism Simulation Environment (DOSE)”, The
Python Papers Source Codes, Vol. 4, 2012, Article 2.
[20] R. W. Hamming, RW, “Error detecting and error correcting
codes”, Bell System Technical Journal, Vol. 29, No. 2, 1950,
pp. 147–160.
[21] T. D. Cuypers, and P. Hogeweg, “Virtual genomes in flux: an
interplay of neutrality and adaptability explains genome
expansion and streamlining”, Genome Biology and Evolution,
Vol. 4, No. 3, 2012, pp. 212-229.
[22] R. E. Lenski, C. Ofria, T. C. Collier, and C. Adami, “Genome
complexity, robustness and genetic interactions in digital
organisms”, Nature, Vol. 400, No. 6745, 1999, pp. 661-664.
[23] J. Sardanyés, S. F. Elena, and R. V. Solé, “Simple
quasispecies models for the survival-of-the-flattest effect:
The role of space”, Journal of Theoretical Biology, Vol. 250,
No. 3, 2008, pp. 560-568.
[24] F. M. Codoñer, J. A. Darós, R. V. Solé, and S. F. Elena,
“The fittest versus the flattest: experimental confirmation of
the quasispecies effect with subviral pathogens”, PLoS
Pathogens, Vol. 2, No. 12, 2006, Article e136.
[25] M. A. Fortuna, L. Zaman, A. P. Wagner, and C. Ofria C,
“Evolving digital ecological networks”, PLoS Computational
Biology, Vol. 9, No. 3, 2013, Article e1002928.
[26] C. D. Pilcher, J. K. Wong, and S. K. Pillai, SK, “Inferring
HIV transmission dynamics from phylogenetic sequence
relationships”, PLoS Medicine, Vol. 5, No. 3, 2008, Article
e69.
[27] A. M. Tsibris, U. Pal, A. L. Schure, R. S. Veazey, K. J.
Kunstman, T. J. Henrich, P. J. Klasse, S. M. Wolinsky, D. R.
Kuritzkes, and J. P. Moore, “SHIV-162P3 infection of rhesus
macaques given maraviroc gel vaginally does not involve
resistant viruses”, PLoS One, Vol. 6, No. 12, 2011, Article
e28047.
[28] D. W. Drury, and M. J. Wade, “Genetic variation and covariation for fitness between intra-population and interpopulation backgrounds in the red flour beetle, Tribolium
castaneum”, Journal of Evolutionary Biology, Vol. 24, No.
1, 2011, pp. 168-176.
[29] L. Raymond, M. Plantegenest, and A. Vialatte, “Migration
and dispersal may drive to high genetic variation and
significant genetic mixing: the case of two agriculturally
important, continental hoverflies (Episyrphus balteatus and
Sphaerophoria scripta)”, Molecular Ecology, Vol. 22, No.
21, 2013, pp. 5329-5339.
[30] J. H. Relethford, “Heterogeneity of long-distance migration
in studies of genetic structure”, Annals of Human Biology,
Vol. 15, No. 1, 1988, pp. 55-63.
50
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Determining Threshold Value and Sharing Secret
Updating Period in MANET
Maryam Zarezadeh1, Mohammad Ali Doostari2 and Hamid Haj Seyyed Javadi3
3
1
Computer Engineering, Shahed University
Tehran, Iran
[email protected]
2
Computer Engineering, Shahed University
Tehran, Iran
[email protected]
Mathematics and Computer Science, Shahed University
Tehran, Iran
[email protected]
secret value periodically update the shared secret is
suggested.
Secret sharing method has many applications in key
management. Li et al. [2] suggested new distributed key
management scheme by combination of certificateless
public key cryptography and threshold cryptography. In
this scheme, for sharing master key of the network, out
of
nodes are chosen as shareholders. Zhu et al. [3],
using threshold cryptography
, presented mobile
agent to exchange topology information and private key.
When a new node requests to connect to network with size
, nodes cooperate and authentication is done. This
method can reduce network overhead and can improve the
success rate of authentication. Zefreh et al. [4] proposed a
distributed CA system based on secret sharing scheme.
They have assumed that the network is divided into several
clusters and each cluster head is in role of distributed CA.
So, a valid certificate is produced by a quorum of cluster
heads. Ge et al. [5] suggested the certificate authority
based on the group of distributed server nodes. In this
model, it is considered different types of nodes jointed to
network and MANET is subject to frequent partitioning
duo to dynamic nature of topology. Hence they classify
nodes to three types: servers, high-end clients and low-end
clients. In requesting procedure, high-end and low-end
clients obtain a valid certification. They sent request to
proxy server and proxy server forwards this request to
other servers. If at least servers exist in group server to
combine at least partial certificates, certificate is issued.
The aim of group key distribution protocol is to distribute
a key used for encrypting the data. Therefore, based on
generalized Chinese remainder theorem, Guo and Change
[6] suggested a group key distribution built on the secret
sharing scheme. Their protocol requires few computation
operations while maintain at least security degree. Liu et al.
[7] proposed similar group key distribution protocol. They
Abstract
In this paper, an attack model is proposed to implement safe and
efficient distributed certificate authority (CA) using secret
sharing method in mobile ad hoc networks (MANETs). We
assume that the attack process is based on a nonhomogeneous
Poisson process. The proposed model is evaluated and an
appropriate amount of threshold and updating period of sharing
secret is suggested. In addition, threshold value effect on security
of the network which uses the distributed CA is investigated. The
results may be also useful in security improvement of networks
that apply secret sharing scheme.
Keywords: Ad Hoc Network, Nonhomogeneous Poisson Process,
Certificate Authority, Public Key Infrastructure.
1. Introduction
Public key infrastructure (PKI) is a basic and fundamental
infrastructure for implementation of security services such
as key generation and distribution in mobile ad hoc
networks (MANETs). In conventional PKI, the centralized
certificate authority (CA) is responsible for the distribution
and management of public key certificate used for
assigning public key to relevant user. But implementation
of PKI in MANET faces several obstacles. Including in
PKI, the conventional single-CA architecture suffers from
single point of failure problem. Furthermore, due to the
dynamic topology and mobile nature of nodes, setting a
node as CA may cause a lot of communication overhead.
The distributed CA method has been suggested for solving
single point of failure problem [1]. In this method, using
threshold secret sharing scheme, functionality of CA is
distributed among several nodes. Then, for providing CA
services, nodes as CA servers cooperate, which is
called the threshold parameter of a secret sharing scheme.
Therefore, the attacker cannot identify the secret key of
CA until it detects the number of sharing secret less than .
Also, to cope with the efforts of attackers to know the
51
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
indicated Guo and Chang ’ s protocol [6] have some
security problems and suggested simpler protocol that
confidentiality of group key is secure unconditionally. In
registration phase, key generation center (KGC) shares a
secret with each group member. Then, KGC establishes
the session key of group using threshold secret sharing
method and Chinese remainder theorem. Each group
member use her/his secret shared with the KGC to recover
the group key. Also, Gan et al. [8] proposed a threshold
public key encryption scheme. In this scheme, on the base
of dual pairing vector space and bilinear group, the
decryption key is distributed between
nodes. For
decrypting the cipher text, it is sent to or more than
nodes and the plain text is obtained, is the threshold
value. For more researches on applications of secret
sharing scheme in ad hoc networks, one can see [9-11].
As can be seen in aforementioned research works, many
studies showed the role of secret sharing scheme in
security of MANET. Then, determination of threshold
value and updating period of sharing secret is important.
However, few studies have focused on this issue. Dong et
al. [12] have compared security of the partially and fully
distributed CA based on the number of server nodes. But
they did not show how to determine the threshold value.
Haibing and Changlun [13] have suggested an attack
model to determine threshold value and updating period of
sharing secret.
The aim of the present paper is to determine the sufficient
threshold value and sharing secret updating period to use
effectively secret sharing scheme in MANETs. For this
purpose, we propose an attack model by considering the
attack process as a nonhomogeneous Poisson process
(NHPP). The paper is organized as follows. Attacks on
MANET and secret sharing scheme are studied in Section
2. Attack process is explored in Section 3. First, some
researches on the attack processes in MANETs are
discussed. Then, in the sequel, the suggested attack model
in this paper is described and a sufficient amount for
threshold and updating period of sharing secret is specified.
In Section 4, the effect of threshold on security of
distributed CA scheme is evaluated. The paper is
concluded in Section 5.
Passive/active: A passive attacker takes an action such as
traffic eavesdropping for information gathering. But in this
attack, no interference is occurred in network host
performance. In active attack, adversary interferes through
actions e.g. modulating, packet forwarding, injecting or
replaying packets, and so on.
Insider/outsider: This is potentially serious security risk
in all security application domains and adversary can
cause with insider capability. Some researchers have
suggested threshold protocols (e.g. m-out-of-n voting
protocols) for resolving this problem in field of secret
sharing and aggregating application protocols.
Static/adaptive: Setting a learning algorithm in each node
can be considered as static. From a practical point of view,
the network's ability to respond to environment, increases
significantly attacker power. For example, make an
informed selection as to which node to compromise next
improves attack performance.
2.2. Attacks on secret sharing scheme in MANET
Since secret is shared between several users and each user
can get only a single secret key, it is difficult to BruteForce attack. It becomes more harder for the adversary to
guess all the values of threshold because value of being
variable for different partitions. But there is a chance of
Brute-Force attack by obtaining partial information from a
shared secret. A malicious user with help of his share of
secret can get another shared secret. When the number of
nodes that is compromised is more than or equal to
threshold, the malicious node can reconstruct secret key.
In other words security of the network has been failed; see
[15]. Yi and Kravets [16] studied the following two active
attacks on a distributed PKI:
(i) Routing Layer Attacks - Malicious nodes disrupt
routing by announcing false routing information such as
injecting incorrect routing packets or dropping packets. If
the attacker blocks or reroutes all victim's packets, some
routing layer attacks can be used to establish a denial-ofservice (DOS) attack.
(ii) Directed Attacks on CA nodes - Once an attacker
discovers identity or location of CA nodes may employ its
resource in attacking only the CA nodes.
2. Attacks
3. Proposed attack model
In this section, we review the attacks on MANET and
secret sharing scheme.
In this section, first, some researches on the attack
processes in MANET are discussed. Then, the suggested
attack model in this paper is described and a sufficient
amount for threshold and updating period of sharing secret
is specified.
Some researchers have studied and modeled the attack
process. Jonsson and Olovsson [17] targeted a distributed
computer system which consisted of a set of 24 SUN ELC
diskless workstations connected to one file server. In
2.1 Attacks on MANET
Many attacks in MANET have been investigated by
researchers. According to [14], attacks on MANET can be
classified as follows:
52
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
(i)
.
(ii) Non-overlapping increments are independent
(iii)
.
(iv)
.
intrusion test is assumed that all attackers are system legal
users with normal user privileges and physical access to all
workstations except file server. They considered intrusion
process into three phases: learning phase, standard attack
and innovative attack phases. Many of data related to
standard attack phase and statistical evidence showed the
intrusion process could be described by an exponential
distribution.
Kaaniche et al. [18] collected data from the honeypot
platforms which deployed on the Internet. Then they did
empirical analysis and statistical modeling of attack
processes. Results showed the probability distribution
corresponding to time between the occurrence of two
consecutive attacks at a given platform can be described
by a mixture distribution combining a Pareto distribution
and an exponential distribution. The probability density
function
, is defined as follows:
where, in little
notation,
when
Interarrival times of Poissson process have exponential
distribution with rate and hence it is said this process has
no memory. This means that
is Poisson
process with mean
where
is
called the mean value function (m.v.f.). NHPP is a
generalization of Poisson process with conditions (i)-(iv)
except for that the rate is a function of denoted by
,
and is called intensity function. The m.v.f of the NHPP is
written as Λ
where
is the distribution function of the time of the first
event in the process. See [20] for a good review of
stochastic processes.
We consider MANET as a closed system. When a network
is based on threshold secret sharing scheme, a group of
nodes have the pieces of secret. Then, for modeling of
attacks, the inside attacks are only considered. Usually in
such attacks, malicious nodes drop and refuse to forward
request of generation or updating certificate, return a fake
reply (e.g. an invalid partial certificate).
It should be noted that
is a probability, is the scale
parameter of the exponential distribution and is the
shape parameter of the Pareto distribution.
It was found that the amount of
in (1) varies from
0.9885 to 0.9981 in all the platforms of honeypot; see [13].
Because
is the weight of exponential distribution in
mixture distribution, Haibing and Changlun [13]
concluded that the exponential distribution dominates
mixture distribution. Also, these authors proposed an
attack model based on Poisson process to determine
threshold value and updating period of sharing secret. In
the suggested model, since the attacks appear according to
a Poisson process then the attacks occur at random instants
of time with an average rate of attacks per second.
Limitation of Poisson process to approximate the attack
process is that its rate is constant and does not vary over
time. On the other hand, it is well known that the security
of MANETs is poor. In other words, the wireless
communication medium is accessible to any entity with
adequate resources and appropriate equipment. Hence,
access to the channel cannot be restricted. According,
attackers are able to eavesdrop on communication and
inject bogus information [19]. This means that over time,
node is more vulnerable to compromise and attack. So, by
considering the rate of attack process as a function of time,
modeling is closer to reality. Based on this, we suggest a
new attack model in which this assumption is also
considered. To describe the model, we need to present the
following definition definition.
Let
be a counting process in which
denotes the number of attacks happened in the interval
. Thus
means that no attack has occurred in
the network up to time
Also, it implies that
denotes the number
of attacks in interval
.The probability that the
network receives attacks in interval
is denoted by:
According to our analysis
can be estimated
by NHPP with m.v.f.
. Then, the probability that
attacks appear during
is obtained as
Particularly, if attacks in the time interval
considered, we can write
is
Due to network protection, all attackers cannot
successfully compromise nodes. We assume that any node
at each attack may be compromised with probability . In
the following, we calculate the probability distribution of
the successful attack process.
Definition: Let
be a non-negative random variable
representing the number of events in the interval
.
Then
is called a counting process. The
Poisson process model is a well known counting process
model. A counting process is a Poisson process if for some
small value h and all times t
Rule 1: Let
be a NHPP with m.v.f.
in
which
denotes the number of attacks happened in the
interval
Assume the probability that the attackers
compromise successfully a node at each attack is and
53
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
called the security of the network and is obtained, as
follows
hence the unsuccessful probability is
. Suppose that
is a random variable denoting the number of nodes
that attackers compromise successfully in the interval
. Then
is a NHPP with m.v.f.
and
hence the probability that nodes are compromised in the
interval
is
is the incomplete gamma
where Γ
function.
Representation (8) shows that the probability of network
security does not depend on the number of nodes. Then,
using
, cannot be compared the security of partial with
full distributed CA schemes.
Also, it can be seen that
is increasing in and is
decreasing in and , separately. Thus, if we select nodes
that are less vulnerable to attacker compromise as nodes
holding the pieces of shared secret in partial distributed
scheme, the security of network is improved.
Using equations (5) and (6), it can be seen that
Hence
for all
which means that no
attack is happened at time
and the network is secure.
The next rule provides the time of maximum compromise
probability.
Rule 2: Consider the assumptions of Rule 1. Further,
suppose that
in which
denotes
the distribution function of the time to the first attack in
the network. The probability that k nodes in the network
have been compromised successfully reaches maximum at
time where
Proof: Differentiating from equation (5), we can obtain
the peak time as follows:
Letting
That is, the maximum value of
depends only on
and is independent of m.v.f. and , which is shown in
figure 1 and figure 2. Figure 1 depicts the 3D plot of
in which
. This plot shows the
independency of maximum value of
from
Also, figure 2 draws
for
and m.v.f.
. As we can see, the maximum value
of
is independent of m.v.f.
, the unique maximum value of equation
(5) is obtained at time
.
Rule 3: Equation (6) will help us to determine the
updating period of sharing secret. Equation (6) states
that when the system is under risk. According to equation
(6) at time attack to nodes as shared secret holders
reaches maximum. In this condition to maintain system
security and prevent of disclosure of shred confidential
should be secret value is updated in
.
.
Rule 4: From (6), it can be concluded that the threshold
value of sharing secret, denoted by . From equation (6),
is obtained. As mentioned until
attacker compromise less than nodes, cannot discover
sharing secret. Therefore, for given attack stream and ,
should be greater than and we can write
4. Evaluation of proposed model
In this section, we will discuss the influence of threshold
value on the security performance of the distributed CA
scheme based on the proposed model. An attacker which
wants to attack the network must compromise no less than
nodes to recover the sharing secret. The probability of
compromising less than nodes in time interval
) is
Fig. 1. The 3D plot
for
.
54
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Fig. 2. The plot
for
been estimated by NHPP. By considering the attack
process as NHPP, the rate of the attack is not necessary
fixed in time as Poisson process and can vary over time.
The results of evaluating of suggested model have shown
that the probability of network security is independent of
the nodes number. Therefore, it cannot be compared the
security of fully distributed and partially distributed CA
schemes using the number of nodes.
According to our analysis, the maximum of probability
that an attacker compromises the nodes of the network
depends only on the number of nodes and is independent
of the m.v.f. This means that if attack process is adopted
based on NHPP, we have a unique maximum for any m.v.f.
Also, the results of network security probability have
shown this probability decreases when the probability that
the attackers successfully compromise a node increases.
Then, the nodes that have less risk of exposure and
vulnerability have been selected as a part of the sharing
secret holder in a partially distributed CA scheme and
hence the security of MANETs can be improved.
.
According to equation (9), the maximum value of
is
decreasing in and is shown in figure 3. In other words,
increasing the number of secret holders, success of
malicious node to obtain secret value decreases. So,
attacker requires to further efforts in fully distributed CA
in compared to the partially distributed CA that all nodes
have a part of secret.
Fig. 3. Relationship between
References
[1] L. Zhou and Z. J. Haas, "Securing ad hoc networks",
Network, IEEE, vol. 13, pp. 24-30, 1999.
[2] L. Liu, Z. Wang, W. Liu and Y. Wang, "A Certificateless key
management scheme in mobile ad hoc networks", 7th
International Conference on Wireless Communications,
Networking and Mobile Computing, IEEE 2011, 1-4.
[3] L. Zhu, Y. Zhang, and L. Feng, "Distributed Key
Management in Ad hoc Network based on Mobile Agent",
in Intelligent Information Technology Application, 2008.
IITA'08. Second International Symposium on, 2008, pp.
600-604.
[4] M. S. Zefreh, A. Fanian, S. M. Sajadieh, M. Berenjkoub,
and P. Khadivi, "A distributed certificate authority and key
establishment protocol for mobile ad hoc networks", in
Advanced Communication Technology, 2008. ICACT 2008.
10th International Conference on, 2008, pp. 1157-1162.
[5] M. Ge, K.-Y. Lam, D. Gollmann, S. L. Chung, C. C. Chang,
and J.-B. Li, "A robust certification service for highly
dynamic MANET in emergency tasks", International Journal
of Communication Systems, vol. 22, pp. 1177-1197, 2009.
[6] C. Guo and C. C. Chang, "An authenticated group key
distribution protocol based on the generalized Chinese
remainder
theorem",
International
Journal
of
Communication Systems, 2012.
[7] Y. Liu, L. Harn, and C.-C. Chang, "An authenticated group
key distribution mechanism using theory of numbers",
International Journal of Communication Systems, 2013.
[8] Y. Gan, L. Wang, L. Wang, P. Pan, and Y. Yang, "Efficient
threshold public key encryption with full security based on
dual pairing vector spaces", International Journal of
Communication Systems, 2013.
[9] S. Yi and R. Kravets, "Key management for heterogeneous
ad hoc wireless networks", in Network Protocols, 2002.
Proceedings. 10th IEEE International Conference on, 2002,
pp. 202-203.
and k.
5. Conclusion
In this paper, we have presented an attack model to
determine the proper value of threshold and updating
period of sharing secret for efficient implementation of
distributed CA based on threshold secret sharing scheme
in MANETs. In the proposed model, attack process has
55
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
[10] J. Zhang and Y. Xu, "Privacy-preserving authentication
protocols with efficient verification in VANETs",
International Journal of Communication Systems, 2013.
[11] A. A. Moamen, H. S. Hamza, and I. A. Saroit, "Secure
multicast routing protocols in mobile ad-hoc networks",
International Journal of Communication Systems, 2013.
[12]Y. Dong, A.-F. Sui, S.-M. Yiu, V. O. K. Li, and L. C. K. Hui,
"Providing distributed certificate authority service in clusterbased mobile ad hoc networks", Computer Communications,
vol. 30, pp. 2442-2452, 2007.
[13] M. Haibing and Z. Changlun, "Security evaluation model for
threshold cryptography applications in MANET", in
Computer Engineering and Technology (ICCET), 2010 2nd
International Conference on, pp. V4-209-V4-213.
[14] J. A. Clark, J. Murdoch, J. McDermid, S. Sen, H. Chivers, O.
Worthington, and P. Rohatgi, "Threat modelling for mobile
ad hoc and sensor networks", in Annual Conference of ITA,
2007, pp. 25-27.
[15] P. Choudhury, A. Banerjee, V. Satyanarayana, and G.
Ghosh, "VSPACE: A New Secret Sharing Scheme Using
Vector Space", in Computer Networks and Information
Technologies: Springer, 2011, pp. 492-497.
[16] S. Yi and R. Kravets, "MOCA: Mobile certificate authority
for wireless ad hoc networks", in 2nd Annual PKI Research
Workshop Program (PKI 03), Gaithersburg, Maryland, 2003,
pp. 3-8.
[17] E. Jonsson and T. Olovsson, "A quantitative model of the
security intrusion process based on attacker behavior",
Software Engineering, IEEE Transactions on, vol. 23, pp.
235-245, 1997.
[18] M. Kaaniche, Y. Deswarte, E. Alata, M. Dacier, and
V. Nicomette, "Empirical analysis and statistical modeling
of attack processes based on honeypots", International
Conference on Dependable Systems and Networks, IEEE
2006; Philadelphia. USA.
[19] J. V. D. Merwe, D. Dawoud, and S. McDonald, "A survey
on peer-to-peer key management for mobile ad hoc
networks", ACM computing surveys (CSUR), vol. 39, p. 1,
2007.
[20] S. Ross, "Stochastic processes", 2nd edn, New York: Wiley,
2008.
56
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Beyond one shot recommendations: The seamless interplay of
environmental parameters and Quality of recommendations
for the best fit list
1
2
Minakshi Gujral1, Satish Chandra2
Department of Computer Science and Engineering, Jaypee Institute of Information Technology,
Noida, UP-201307, INDIA.
[email protected]
Department of Computer Science and Engineering, Jaypee Institute of Information Technology,
Noida, UP-201307, INDIA.
[email protected]
Recommendation Systems lays strong foundation for the
commercially rich and ever expanding Recommendation
culture with ever growing online commerce. Not only it
proves solution for Information Overload problems but
also provides users with novel and preferred products,
concepts and services relevant to their preferences. The
Social aspect of Recommendations shapes the cultural
flow. As our culture moves online, the creativity,
evolution and augmentation of connection centric
recommendation process grows four folds. Rightly said by
Fortune magazine writer Jeffrey M. O'Brien, in an article
published in CNN Money, entitled “The race to create
'smart' Google”,”The Web, they say, is leaving the era of
search and entering one of discovery. What's the
difference? Search is what you do when you're looking for
something. Discovery is when something wonderful that
you didn't know existed, or didn't know how to ask for,
finds you”. The quest for best fit decisions is about
moving ahead of search engines in the world of
recommendation.
This proposed solution analyzes some selected but generic
recommendation systems of varied application
environments, and their recommendation methods,
performance, user and item profile, rating structures,
similarity measure, and other issues. This gives a Multi
dimensional evaluation framework to model optimized
system for the best fit recommendations. This kind of
approach has cropped from recommendation system’s
evaluation and research for optimized recommendations.
It started from way back mid 1990 to this current era but
no concrete, balanced and feasible solution from
multidimensional perspective has been given till date.
This analysis can be extended with no of issues, systems
and their changing context. There are viable paths of
improvements and extensions, which can be implemented,
mixed and matched for feasible, environmentally tailor
made best fit recommendations.
Abstract
The Knowledge discovery tools and techniques are used in an
increasing number of scientific and commercial areas for the
analysis and knowledge processing of voluminous Information.
Recommendation systems are also one of Knowledge Discovery
from databases techniques, which discovers best fit information for
appropriate context. This new rage in Information technology is
seen in area of E-commerce, E-Learning, and Bioinformatics,
Media and Entertainment, electronics and telecommunications and
other application domains as well. Academics, Research and
Industry are contributing into best-fit recommendation process
enrichment, thereby making it better and improvised with growing
years. Also one can explore in depth for qualitative and
quantitative analysis of E-World Demand and Supply chain with
help of recommendation systems. Lot has been talked about
effective, accurate and well balanced recommendations but many
shortcomings of the proposed solutions have come into picture.
This Paper tries to elucidate and model Best Fit Recommendation
issues from multidimensional, multi-criteria and real world’s
perspectives. This Framework is Quality Assurance process for
recommendation systems, enhancing the recommendation quality.
The proposed solution is looking at various dimensions of the
architecture, the domain, and the issues with respect to
environmental parameters. Our goal is to evaluate
Recommendation Systems and unveil their issues in quest for the
Best Fit Decisions for any application domain and context.
Keywords: Recommendation Systems, Best fit decisions, Issues
in Recommendations,
decisions.
Expert
recommendations,
best fit
1. Introduction
The Recommendation systems are the new search
paradigm which has stormed the E-world with semantic
preferences and excellent information services. Some of
the renowned recommendation systems are Amazon,
last.fm, Netflix, Cinematch, yahoo and Google.
57
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
large scale application domains like recommending
vacations, financial services. The flipside is although
these Extensions are introduced but till date neither
implemented nor explained concretely and explicitly in
recommendation system’s research. Also the changing
context
and
knowledge level influences the
recommendation results.
The recommendation Quality has multilingual aspects to
it while diving deep in research, innovation and novel
ideas for recommendations. It is the search for best fit
solution for any application, technology, for varying
demography of users. The need is to explore Multi
dimensional issues for Multi criteria parameters to
produce best fit optimized recommendations. This paper
tried to ponder on best fit recommendation issues in
section 2, evaluates it in section 3 and gives a knowledge
prototype approach in section 4.
2.2 Issues hindering best fit Recommendations:
The study of various recommendation issues in this
scenario gives a new dimension through some
formulations. Firstly the various unresolved extensions
introduced:
a) Comprehensive Understanding of Users and Items
b) Model Based recommending techniques
c) Multidimensionality of Recommendations
d) Multi-criteria Rating
e) Non-intrusiveness
f) Flexibility
g) Trustworthiness
h) Scalability
i) Limited Content Analysis
j) Over specialization
k) New Item/User Problem
l) Sparsity
2. Exploring Multi Dimensional Issues in
Recommendation Systems [1-4]:
2.1 Evaluating Recommendation system’s algorithm:
The evaluation of various recommender systems’
algorithm is done for jargoning and validating the issues.
Thereby the focus is on two kinds of evaluations:
1) The first one concerns the performance accuracy with
respect to changing context and knowledge levels.
2) The second concerns user satisfaction, retaining the
interest of users and formulating the requirement model
of the problem task.
Several approaches are compared by the tool used in
experiments on two real movies rating datasets i.e. Movie
Lens and Netflix. The collaborative and content filtering
algorithms, used in recommender system, are
complementary approaches. Thus this motivates the
design and implementation of hybrid systems. And thus
this new hybrid system is tested with real users.
Various extensions in recommendation capabilities are
rightly focused but not justified because:
1) Problems and Extensions are introduced theoretically
but not yet solved from multi dimensional real world
scenario. The thematic profiles of Users and Product
Attributes are evaluated and updated theoretically with
synthetic data sets but real life transactions give the clear
picture. These formulations need further validation with
effective feedback of real world data.
2) Introduction of Contextual, critiquing and
Conversational Recommendations should be incorporated.
3) These extensions should be balanced with increase in
information, user, network and addition of complex cross
application networks.
Following deductions give some focal points to this
approach:
1) Similarity Measure cannot be implemented for all users
of varying preferences.
2) An unresolved issue is to explore criteria that try to
capture quality and usefulness of recommendations from
the user satisfaction perspectives like coverage,
algorithmic complexity, scalability, novelty, confidence
and trust. This all is important along with user-friendly
interaction.
3) The need to design an efficient, user friendly user
interface which keeps the user from quitting and
extracting important information for making requirement
model of the problem task.
4) Traditional Recommendation Methods i.e. content,
collaborative and hybrid [depending on rating system],
with their advantages and shortcomings contribute
towards Possible extensions. This capacitates them for
2.3 Background study of some recommendation
systems:
1) An innovative approach for developing reading
material recommendation systems is by eliciting domain
knowledge from multiple experts. To evaluate the
effectiveness of the approach, an article recommendation
expert system was developed and applied to an online
English course.
But the flipside measured in this case: Learning
preferences or needs are not the same. Online e-learning
58
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
computational cost increases. Practical problems arise in
this case i.e.
-New user new item problem
-Cold Start Problem
-Coverage Issues
-Computational Cost
6) Some recommendation research also tried to explore
Singular Value Decomposition (SVD) to reduce the
dimensionality of recommender system databases to
increase performance of recommendations [6]
a) SVD can be very expensive and is suitable for offline
transactions. Its complex architecture, differentiating
results for different setups also hinders good
recommendation approach. This does not take into
account the Top N Predictions in a cot effective way to
enhance
quality
of
recommendations.
Here
Recommendation Evaluation speaks of more issues to
ponder. For example Best Fit Recommendations in
changing E-WORLD, security issues, measuring user
satisfaction,
efficiency,
and
persuasiveness
of
recommendation process, scalability, coverage, and these
points are not fulfilled in such evaluations.
b) The Concept of [Rating, Prediction Elicitation, and
Similarity Measure] is vague and difficult in practical
scenario. They are taken from synthetic Datasets. Real
world transactions say reservation system, registration is
far better than synthetic datasets; this approach fails to
implement it. The thematic profiles of Users and Product
Attributes are implemented theoretically but not from real
life transactions. These calculations further needs to be
checked with effective feedback of real world data.
c) Overhead of Computation of Similarity Measures is
also there. Out of Box thinking is needed to measure
Similarity not only in the thematic profiles of Users and
Product Attributes theoretically but also from real life
transactions. These calculations further needs to be
checked with effective feedback of real world data.
module provides fixed learning content for all students
with varying aptitude and knowledge level.
2) Technique using the repertory grid method assists the
domain experts to better organize their knowledge and
experiences. This is significant approach but with ever
changing and evolving nature of user and item model
there arises doubts on this approach being successful. It’s
difficult to calculate recommendation best for different
learning levels. Other factors like Authenticity of data
filled, coverage of application domain with increase in
courses and students, absence of Justification of choice by
users also contribute to this discussion and points towards
Multi dimensional issues.
3) Another novel research problem which focuses on the
personalization issue is the problem of recommendation
under the circumstance of a blog system. It is a
comprehensive investigation on recommendation for
potential bloggers who have not provided personalized
information in blog system. This talks about registered
and unregistered bloggers (personalized/non personalized)
given recommendations for services in blog environment.
4) Another Recommendation Method presents an
algorithm on the inhomogeneous graph to calculate the
important value of an object, and then combined with
similarity value to do recommendation.
This is a random work in inhomogeneous graph thereby
this cannot be generalized with increasing users and
resources. The following points come in front:
a) There is no information about what happens to
recommendations with new users and resources.
b) There is no clear picture about how one can calculate
best recommendations for a particular item type which
does not match neighbor nodes?
c) The issues like scalability and trustworthiness of the
system are at stake. Normally any blogger can register
and can create imperfect/illegal data or change
information which is again one contradictory point.
d) Furthermore the methodology is too voluminous,
complex to implement for large application domains.
5) Some recommendation algorithm tried to incorporate
“thinking of out of the box recommendations” [5]. This
Concept introduces TANGENT, a novel recommendation
algorithm, which makes reasonable, yet surprising,
horizon-broadening recommendations. The main idea
behind TANGENT is to envision the problem as node
selection on a graph, giving high scores to nodes that are
well connected to the older choices, and at the same time
well connected to unrelated choices.
Extensive comparison of the numerous alternatives; while
all seem reasonable on paper, several of them suffers from
subtle issues. Computational cost being one of them. Also
at some undesired situation, if neighbor’s nodes are at
complex
positions,
complexity,
overhead
and
3. Evaluating multidimensional issues of
recommendation process:
3.1. The Cold Start Problem
The cold start problem occurring when a new user has not
provided any ratings yet or a new item has not yet
received any rating from the users. The system lacks data
to produce appropriate recommendations. [1][2].To
remove over specialization problem when there is use of
diversification in explanations [11] this gives chance to
new items in the group which is good testament to
solution of cold start problem. CBR plus Ontology [12]
59
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
need logic or reasoning to calculate satisfaction which is
an inexplicable concept.
concepts reasons out the current and old status of items as
per logic based calculations rather than ratings. This
increases the system complexity and even if it gives cold
start solution to 15% it increases scalability sparsity and
other issues. This 15% also is viable when information is
well fed by past cases. Knowledge based Models[6] hits
hard on usage of rating structure by Implicit
recommendation methods[1,2] and proposes evaluating
explicit user/item models to take care of cold start
problem. The excellent analysis done by knowledge
Model framework[16] which says that as new products of
various categories arrive in market, this can further ignite
cold start, over specialization and sparsity problem. Even
clubbing Intelligent Agents and mining Semantic web [18]
leads to cold start problem and with increase in scalability
the system suffers. The Cold start problem is apparent in
systems depicted by Table 1.
3.4. User Satisfaction Issue
An important issue in recommender system now is to
explore criteria that try to capture quality and usefulness
of recommendations from the user’s satisfaction
perspectives like coverage, algorithmic complexity,
scalability, novelty, confidence and trust, user interaction.
The need to design an efficient, user-friendly user
interface: [1-3, 5]
1) For Mixed hybrid approach to implement again there is
a decision i.e. which items should be rated to optimize the
accuracy of collaborative filtering systems, and which
item attributes are more critical for optimal content-based
recommendations, are issues that are worth exploring.
2) Even if the recommender system is accurate in its
predictions, it can suffer from the ‘one-visit’ problem, if
users become frustrated that they cannot express
particular preferences or find that the system lacks
flexibility. Creating a fun and enduring interaction
experience is as essential as making good
recommendations.
Moreover Table 1 depicts this problem in some systems.
3.2. Coverage
While preferences can be used to improve
recommendations, they suffer from certain drawbacks, the
most important of these being their limited coverage. The
coverage of a preference is directly related to the coverage
of the attribute(s) to which it is applied. An attribute has a
high coverage when it appears in many items and a low
coverage when it appears in few items [1] [2]. Table 1
depicts the coverage problems.
3.5. Personalization
Recommendations.
3.3. Overspecialization
as
potent
factor
in
As depicted by Table 1, personalization is a potent factor
to be solved:
Processing of User Preference is Difficult due to:
1) Different User Background
2) Registered/Unregistered Users (TRUST ISSUE)
3) Dynamic Remodeled Information.
4) Willingness/Decision making criteria of user for best
fit preference.
5) Personalization should be clubbed with security issues.
Content-based approaches can also suffer from
overspecialization. That is, they often recommend items
with similar content to that of the items already
considered, which can lead to a lack of originality. [5][6].
Moreover Table 1 depicts the Overspecialization problems
in some systems. As per Conversational strategy also
which is best of above mentioned lot, all the preferences
to be specified upfront, at the beginning of the interaction.
This is a stringent requirement, because users might not
have a clear idea of their preferences at that point. New
Preferences and servicing needs to be feed in, system itself
cannot predict or recommend and again presents with old
strategies and products [14]. Overspecialization is said to
be solved effectively by attribute based diversity and
explanation based diversity[11], of which explanation
based diversity strategy reigns supreme as it takes care of
computational cost, provides coverage for even social
domain, and is better in efficiency and performance. But
needlessly explanation is also criteria added to content or
collaborative based technologies, so it cannot escape from
its structural disadvantages. Furthermore explanations
3.6. Scalability
Some of the systems have limitation of scalability as
depicted by Table 1. By increasing load on the
recommendation in terms of growing item, users, the
system slows down effective process. This degrades
system performance, efficiency and throughput of
recommendation system. In using Explanation Facility to
solve over specialization[13] and to bring in diversity in
product choice[11], the recommendation quality improve
but with increase in users, items and modules of system,
the result is complexity and overhead which further
breaks the performance. Same happens with other systems.
60
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
it profitable to shill recommender. The Affiliate
marketing and other E-Commerce profit gain concepts
also fall in this category. This all is done to have their
products recommended more often than those of their
competitors. Such acts are known as shilling attack on
recommendations. Also the concepts from a well known
recommendation work on research domain elucidates
some unwanted situations and the framework for shilling
attack as a recommendation issue and the respective
formulations given below. This work defines a new issue
named shilling attack on quality of recommendation on
the basis given below. Method Citation is considered
equivalent to rating phenomenon. Four Formulations [F1F4] is given on these grounds:
a) Matthew Effect describes the fact that frequently cited
publications are more likely to be cited just because the
author believes that well-known papers should be
included.
F1: Branded/well known items are given more
preference/rating not even knowing that they match the
demography of users and context of his query. This can
further lead to Over specialization, cold start and sparsity.
For example any new shoe category branded X Company
has to be rated 8/10, without actually checking out the
product. This can be termed as RecommBrandedRating
Effect.
b) Self citations: Sometimes self citations are made to
promote other publications of the author, although they
are irrelevant for the citing publication.
F2: At times when items are recommended, there are
some complementary things which are presented along
with that, which may not match the preference elicitation
or context. This is just done for gaining profit from other
company/advertising. This can lead to Computational cost,
decay in performance, efficiency, user satisfaction,
coverage and personalization issues. Ratings/Preferences
or recommendation results are more biased towards own
company product thereby to increase brand value. This
can be termed as RecommSelfRating Effect.
c) Citation circles: In some cases Citation circles occur if
citations were made to promote the work of others,
although they are pointless.
F3: Group Recommendations following F1 and F2 can be
construed in this category. This can further ignite sparcity,
cold start, and overspecialization. Even the corporate
world revert their direction towards ‘give and take’
phenomenon where other products are rated high, who in
turn favors you with high recommendation value of home
products. This can be termed as RecommAffiliateRating
Effect.
d) Ceremonial citations: The Ceremonial citations are
citations that were used although the author did not read
the cited publications.
Increase in load or scalability is real test of system
potential and capacity. Research solutions on small scale
or fewer loads are feasible from all perspective but with
scalability of system it is a different scenario in itself.
3.7. Scrutability
There are Recommendation Frameworks which exhibit
absence of scrutability criteria as per Table 1. Scrutability
is one of recommendation quality parameter which
permits user to alter his query to tailor fit his
recommendations. [13] The user is giving his authentic
feedback for the recommendations via scrutability. Many
recommendation process do not allow user feedback or
scrutability. This can lead to dissatisfaction of user.
3.8. Sparsity
There are Recommendation Frameworks, in which there
are sparsity issues as shown by Table 1. The Concept of
sparsity leads to a situation when enough transactional
data is not there for any type of correlation or mapping of
[item/user] data. Be it recommendation technique of
recommendation using diversification [11], explanation
facility[13], tags[15] and others, most of them lack
transactional, linking data to map or correlate user/item
models. In other words calculating the distance between
entities [user/item] becomes difficult.
3.9. Introducing the concept Recommendation
shilling attack [7, 8]
This is a Novel Contribution, which is a part of the
proposed Solution. But the background research for this
formulation is given by study of some of such issues
coming from fake ratings systems of research citations
and E-Commerce world.
E-Commerce Security Issues like “shilling attack” can
also be found in well known research context issues like
Matthew Effect, self citations, citation circles and
ceremonial citations.
Recommendations in ratings can be faked just as
publication numbers, citations and references are inserted
for good results. Ratings can be added to high numbers of
desired products, and even decreased. There arises a need
for semantic structuring and authenticity.
As shown in Table 1, some of the Systems suffer from
Recommendation Shilling attack. Unscrupulous producers
in the never-ending quest for market penetration may find
61
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
extent and Explanation facility, further augmented this
approach. The Concept of Search Algorithms, repertory
grids and Graph models also participated to churn out
optimized best fit recommendations. Conversational,
Context and Critiquing process brought a see saw change
in recommendation quality research. Even Case Base
reasoning, Ontology, Intelligent Agents also contributed
towards this direction. The innovation and feasible
technique illuminated by knowledge based models is a
show stopper. It models.
The Learning Techniques are described by the strategy of
recommendation process. Various systems used in Table 1
have following learning Techniques.
a) Traditional systems based on [1, 2] Rating structures,
User/item profile analysis [Content/ Collaborative/hybrid]
b) Personalization-User Demography [19]
c) Search Algorithms
d) Tagging Concept in Recommendations [15]
e) Explanation Facility [11, 13]
f) Knowledge Models [4, 16, 22]
g) Repertory Grids [3]
h) Graphs [6, 9, 10]
i) Capturing Context through Conversational/Critiquing
Strategies [14, 23]
j) Capturing Context using semantic web [17]
k) Using CBR+ONTOLOGY Combination [2]
l) Using intelligent agents for Recommendations [18, 21]
m) Blogs/Social Groups based Connection Centric
Recommendations.
Example: Orangut/twitter/LinkedIn/Wayn [9]
F4: At times raters or recommendations are just listed for
increased brand name in market without Comprehensive
understanding of user and items. Without calculating the
similarity
measure/personalization/preference
of
recommendations some items are rated arbitrarily.
All Four factors affect recommendation quality, even
discourage future extensions. This intensely affects the
comprehension understanding of users and Items and also
inversely affects User Satisfaction level. This can be
termed as RecommAstractRating Effect.
This Four Formulations has been grouped under the
concept Recommendation shilling attack. This has been
applied to concept of shilling attack with perspective of
quality framework. One can well imagine if such fake
entries enter in recommendation process, what will be the
overall effect on item preferences, cold start, over
specialization and other issues. It would be disastrous.
3.10.
The
Central
Processing
Unit
Recommendation
Framework:
Quality
Recommendations. [1-3, 6, 9, 10-17, 21, 23]
of
of
The Quality of Recommendations is measured by some
primary component such as Novelty, flexibility,
scrutability, transparency, effectiveness, efficiency,
persuasiveness. But presence of issues depicted by Table 1
fails the most tailor made recommendations also. So the
need arises to evaluate the recommendation quality from
Multi dimensional qualitative perspectives as well. These
are given by primary components of recommendation
quality parameters as described above. The various
learning techniques i.e. ontology, repertory grid etc of
these algorithms are evolved to resolve recommendation
issues. Recommendation strategies also encompasses
evaluations grounded in mathematical logic such as
Pearson’s co-efficient, top N Model, but they are
algorithmically striking a balance among prediction,
recommendation, relevance, diversity of items and
measuring of issues which is not viable. This requires
feasible real world modeled framework.
The recommendation Quality is much more than these
seven primary factors. They are affected by 12 issues of
Table 1 framework. Quality needs to be seen from
multidimensional perspectives.
Quality of recommendations depends on many factors;
1) Recommendation Issues.
2) The Learning Techniques and the variables used in it.
3) The real time evolution of Application Framework.
Therefore the 12 factors of Table 1 have quality
parameters embedded in them.
3.11 Environmental parameters of Recommendation
Evaluation. [25-27]
The quality of Recommendation is a Multi dimensional
evaluation from all these twelve perspective. The quality
evolution started from content, collaborative and hybrid
techniques. Ratings helped to calculate user satisfaction,
Personalization brought semantic understanding to it.
Recommendation Quality research further saw similarity
measure, preference elicitation, difference between search
recommend and predict to ascertain the exact co-relation
between user characteristics and item attributes. Tagging
facility solved the understanding of item background to an
Issues like Multidimensionality of Recommendation
system, Multi criteria Rating and Model based
Recommendation techniques calls for a Frame work
which takes information from user with user friendly GUI
in conversational stages/increments. This further
processes contextual ratings using KBB technique from
62
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
in
Recommendation density, consumer taste, and consumer
taste variability and consumer type.
This Concept refers to Recommendations which fits in
the Multi Dimensional Criteria’s of user and application
domain. For example demography of user, object
attributes, social, legal and environmental rule based
structure, understandable and cost-effective navigational
links. It has Multi-User Flexible Interface to cater to user
needs i.e. stop the user from one visit problems and assist
him to decide. This further assists to create homogenous
environment of interaction. But some of the
Recommendation systems lacks in one way or the other in
this type of evaluation point. As depicted by Table 1. In
this work, the novel contribution is the 3M Model which
has been identified from the above stated concept. This is
of Multidimensional Multi-criteria modeling: The 3MModel.
From the perspective of E-Business point of view:
Recommendation System’s Maintenance Model [Business
Model]: This deals with business model perspective of
recommendations thereby taking care of real-time
transactions and the cost incurred.
a) Maintenance Model 1: charge recipients of
recommendations either through subscriptions or pay-peruse [Danger of fraud].
b) Maintenance Model 2: A second model for cost
recovery is advertiser support. [Danger of fraud]
c) Maintenance Model 3: A third model is to charge a fee
to the owners of the items being evaluated. [Partial
Danger]. This 3M Model can also take care of TRUST
Issues.
This has given a balanced structure to evaluate and satisfy
various recommendation parameters. Therefore following
analysis holds true:
a) Recommendation Quality Framework: Good Evaluation
Matrix which enhances recommendation quality. It talks
about the parameters of quality evaluations which can
provide sound base to explain issues. For example:
Recommenders Density of recommendations tells you
about over all coverage or sparsity criteria’s.
b) It also evaluates the Cost structure therefore evaluating
the remuneration of each approach and strategy.
c) Consumer Taste parameter evaluates the over
specialization issue also to some extent can help to solve
collaborative
filtering
disadvantages
due
to
inconsideration of different user background.
d) Recommendation system’s implementation, usage and
evaluation are a costly transaction. This has to be
strategically managed by structuring balanced quotient of
multidimensional Recommendation quality criteria’s and
Business Model’s needs. Cost and quality parameters
should be managed according to application domain’s
need and user’s perspective.
e) Qualative parameters alone cannot justify best fit
recommendations; they need to work with issues. They
need justification for analysis and an architectural
framework for generation of recommendations.
E-COMMERCE
knowledge Grid.
Databases
thereby
processing
We can look at recommendation system from social and
technical evaluation perspective. For this the author has
taken 5 recommender systems Grouplens, Fab, Referral
web, PHOAKS AND SITECEER.
This evaluation metrics further illuminates various
recommendation methodologies, architecture and their
implementation and processing strategies with two main
goals:
1) Recommendation system’s quality evaluations
2) Recommendation system’s maintenance in real world.
For this the evaluation metrics are categorically divided
into three major parts of evaluations:
The Technical Design Space: This consists of Contents
of recommendation, use of recommendations, explicit
entry, anonymous and aggregation,
The domain space and characteristics of items
evaluated: This consists of type, quantity, and lifetime
and cost structure of Recommendations.
The domain space and characteristics of the
participants and the set of evaluations: This consists of
ISSUES
S1
S2
S3
S4
S5
S6
S7
S8
S9
S10
S11
S12
Accurate
Rating
1
1
0.5
1
1
1
1
1
1
0.5
1
1
Preference
Elicitation
1
1
0.5
1
1
1
1
1
1
0.5
1
1
Similarity
Measure
1
1
0.5
1
1
1
1
1
1
0.5
1
1
Building
of item/
user profile
1
1
0
1
0.5
1
0
1
1
1
1
1
Personalized
recommendation
1
1
0
1
0.5
1
0
1
1
1
1
1
1
1
0
1
0.5
1
0
1
1
1
1
1
1
0
0
1
1
1
1
0
0
0
0
0
Implicit/
Explicit Data
Collection
Cold Start
Problem
Coverage
0
0
0
1
0
1
1
0
0
0
0
1
Sparsity
0
0
0
1
0
1
1
1
1
0
0
1
Over
specialization
1
0
0.5
0
1
0
1
1
0
0
0
1
Recommend
-ation
Quality
1
1
0.5
1
0
1
1
1
0.5
0
1
1
User
Satisfaction/1
Visit Problem
1
0
0.5
0
0
1
1
1
0
1
0.5
1
63
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Scrutability
1
1
1
1
0
0
1
1
0
0
1
1
Scalability
0
0
0
0
0
0
1
0
0
0
0
1
Real World
Modeling
0
0
0
0
0
1
1
1
0
0
1
1
Shilling Attack
0
0
0
0
0
0
0
0
0
1
0
0
Personalization- 05
Scalability- 06
Scrutability- 07
Sparsity- 08
Real World Modeling of data. - 09
Layer 3: The Recommendation Density and other
Quantitative parameters are referred as:
Function F= {A1…..AN} where N is all the quantitative
parameters. This Function F is further checked with the
information overload problem due to social networking.
This is done with the viewpoint of Social Centric View of
World Wide Web i.e. harnessing the Recommendation
linkages from social centric network [25][26][27]. This
step is important to recover from problems of
Personalization, Navigational links and Social Networks.
Table 1: Multi Dimensional Issues in Recommendation Systems Matrix. Here
Systems S1..S12 refers to systems and concepts depicted by research papers
[11-22] at back reference in same order. The number 0 marks presence of
issue. 1 means system is working fine with respect to following issue. The
table can be read as: a) Coverage Problems are there in some systems [11-13,
15, 18-21]. b) Overspecialization problems are there in some systems [12, 14,
16, 19, 20, and 21]. c) User Satisfaction Issue are prevalent in systems [12,
14, 15, 19].d) Personalization problem is there in systems [13, 17].e)
Scalability is there in some of the systems [11-16, 18-21]. f) Scrutability
problem present in systems [15, 16, 19, 20]. g) Sparsity is shown by some of
the Recommendation Frameworks [11, 12, 13, 15, 20, and 21]. h) As shown
in Table 1, some of the systems [11-19, 21-22] also suffer from
Recommendation Shilling attack. i) System in References [12, 13, 18-22]
depicts Cold start problem. Real world Modeling problem is shown by the
systems [1-5, 9, 10].
Layer 4 and 5:
For system S1 the tags allocated going through various
layers starting from Layer 1 to Layer 5.
To solve the probability of Recommendation shilling
attacks, tags are further extended by adding extensions
F1..F4
{Formulations
discussed
in
section
Recommendation shilling attack} -> For S1 We have
S1%26789%F2F3.
4. The Recommendation evaluation
framework by Multi Layer Architecture:
This Evaluation prototype paves the way for measuring
the recommendation process of best fit recommendations.
This is a generic, abstract framework which can further be
extended by increasing number of systems, architecture,
issues, measurement criteria’s and viable solution sets.
This can be further given the shape of Knowledge Grid.
This Consists of 8 layers and 3 Knowledge Bases.
Referring to Table 1, the working of the framework is
explained with respect to system S1.
Layer 6: Here we resolve the fake rating structure or data
due to recommendation shilling attack. We remove the
redundant data added due to fake ratings thereby adding
Clear Flag giving S1%26789%Clr.
Layer 7: further matches the Solution set Knowledgebase
SKB3 which consists of strategies to resolve issues. This
is future work and this may give rise to many other
research directions. For system S1 we have to resolve
issues 2, 6, 8, 9. This may further give solution sets in
terms of architectural, environmental, qualitative or
quantitative changes to rectify the given problems.
Layer 1: Extracts all the Algorithmic Detail of system S1
i.e. Architecture {Knowledge Model, Ontology, Repertory
Grid etc}, Domain Application {E-Commerce, Elearning}, Quantitative Parameters [25][26] {Number,
type, Cost and lifetime of recommended items, Contents
and density of Recommendation, User taste variability etc}
and store in Knowledge Base as KB1.
Layer 8: tests the validity of the recommendations. In
case of failure, it sends this to layer 1 or exits the system.
Each Recommendation Data goes through maturation
effect i.e. data change due to environment or knowledge
level of people so layer 1 again tracks all the system
information. Thereby it is customary to update KB1, KB2
and SKB3 with evolution in recommendation data and
strategies.
Layer 2: The Issues are recorded in a grid framework
with severity of issues mapped as {X=0, √=1, 1/2√
=50%}. The Issues are given a Tag which consists of
Issue Number. The Issues are stored in Knowledge Base
as KB2, according to their sequence Number and Issue
details.
The Cold Start Problem - 01
Coverage - 02
Overspecialization - 03
User Satisfaction Issue - 04
5. Conclusions
Recommendation System can be termed as Information
Filtering KDD Technique which evolves with E-world. It
64
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
be done and more importantly issues have to be clearly
and broadly classified and worked out for approximate,
contextually best fit decision.
is a way ahead of smart search. Every Recommendation
Algorithm in spite of anomalies have basic Learning
Technique, Issues creating problems in good
recommendations,
parameter
to
evaluate
recommendations, social centric point of view, all
contributing to quality framework for best fit
recommendations.
References
[1] L. Candillier, K. Jack, F. Fessant, and F. Meyer, "State-of-theArt Recommender Systems," in Collaborative and Social
Information Retrieval and Access: Techniques for Improved
User Modeling. M. U. o. T. Chevalier, Ed. 2009, pp. 1-22.
[2] Adomavicius, G., & Tuzhilin, A. (June 2005). Toward the
next generation of recommender systems: A survey of the
state-of-the-art and possible extensions. IEEE Transactions on
Knowledge and Data Engineering, Vol 17,Issue 06,pp 734–
749.
[3] Hsu, C., Chang, C., and Hwang, G. 2009. Development of a
Reading Material Recommendation System Based on a Multiexpert Knowledge Acquisition Approach. In Proceedings of
the 2009 Ninth IEEE international Conference on Advanced
Learning Technologies - Volume 00 (July 15 - 17, 2009).
ICALT. IEEE Computer Society, Washington, DC, 273-277.
[4] Hsu, C., Chang, C., and Hwang, G. 2009. Development of a
Reading Material Recommendation System Based on a Multiexpert Knowledge Acquisition Approach. In Proceedings of
the 2009 Ninth IEEE international Conference on Advanced
Learning Technologies - Volume 00 (July 15 - 17, 2009).
ICALT. IEEE Computer Society, Washington, DC, 273-277.
[5] Onuma, K., Tong, H., and Faloutsos, C. 2009. TANGENT: a
novel, 'Surprise me', recommendation algorithm. In
Proceedings of the 15th ACM SIGKDD international
Conference on Knowledge Discovery and Data Mining (Paris,
France, June 28 - July 01, 2009). KDD '09. ACM, New
York, NY, pp 657-666.
[6] B. M. Sarwar, G. Karypis, J. A. Konstan, and J. T. Riedl,
"Application of dimensionality reduction in recommender
systems–a case study," in In ACM WebKDD Workshop,
2000.
[7] B. Gipp and J. Beel, “Scienstein: A Research Paper
Recommender System,” in International Conference on
Emerging Trends in Computing. IEEE, 2009, pp. 309–315.
[8] S. K. Lam and J. Riedl. Shilling recommender systems for fun
and profit. In WWW ’04: Proceedings of the 13th
international conference on World Wide Web, pages 393–402,
New York, NY, USA, 2004. ACM Press.
[9] Perugini, S., Gonçalves, M. A., and Fox, E. A. 2004.
Recommender Systems Research: A Connection-Centric
Survey. J. Intell. Inf. Syst. 23, 2 (Sep. 2004), pp 107-143.
[10] Z. Huang, W. Chung, and H. Chen, "A graph model for ecommerce recommender systems," Journal of the American
Society for information science and technology, vol. 55, no. 3,
pp. 259-274, 2004.
[11] Yu, C., Lakshmanan, L. V., and Amer-Yahia, S. 2009.
Recommendation Diversification Using Explanations. In
Proceedings of the 2009 IEEE international Conference on
Data Engineering (March 29 - April 02, 2009). ICDE. IEEE
Computer Society, Washington, DC, 1299-1302.
[12] Garrido, J.L.; Hurtado, M.V.; Noguera, M.; Zurita,
J.M.,"Using a CBR Approach Based on Ontologies for
Recommendation and Reuse of Knowledge Sharing in
This paper categorizes recommender systems semantically
on large scale. This includes applications ranging from ecommerce to social networking, platforms from web to
mobile, healthcare to differential diagnosis, project
management to quality assurance and beyond, and a wide
variety of technologies ranging from collaborative
filtering, content based, hybrid approaches, explanations
facility, tags terminology, ontology, case-based reasoning
to Knowledge Models.
Thus a need arises to evaluate and explore
recommendation architecture with perspective of issues,
quality parameters of evaluation.
The main goal is trust worthy, user friendly,
conversational, context enriched, novel best fit
recommendations. This all contributes to Multi
dimensional evaluation architecture which filters cost
effective,
application/domain
based
best
fit
recommendations. The summary Table 1 encompasses
semantic analysis of some selected systems and can be
extended with n number of systems and other issues as
well. The era of best fit recommendations tailor made for
any application domain will see the user asking for,” what
you suggest for me? Rather than, “I am X ,I have used Y,
and can you suggest something like Y?.
Further Recommendation system needs more cross
dimensional analysis from the perspective of: Issues,
Qualitative and Quantitative Framework, Business Model,
Architecture and Application Domains.
In this Paper the evaluation frameworks of very few
architecture with respect to issues and quality parameters
have been evaluated. The prototype of evaluation is
present but this consists of core areas of research, precise
measurements and boundary definition which is a future
extension. The main focus of this paper has mainly been
discussion of issues and trying to eradicate them by giving
an algorithm for filtering the issues through tags. This
algorithm needs to be extended to a bigger architecture
with respective functions simulated with datasets. There
are many Issues and some of them are overlapping as well.
The accurate and best fit recommendation generating
evaluation framework is future work. There is lot more to
65
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
[26] Gujral, M. and Asawa, K., Recommendation Systems - The
Knowledge Engineering analysis for the best fit decisions ,
Second International Conference on Advances in Computer
Engineering – ACE 2011, Trivandrum, Kerala, INDIA.
Decision Making",Hybrid Intelligent Systems, 2008. HIS '08.
Eighth International Conference on 10-12 Sept. 2008,
Page(s):837 - 842.
[13] Tintarev, N.; Masthoff, J.,"A Survey of Explanations in
Recommender Systems",Data Engineering Workshop, 2007
IEEE 23rd International Conference on 17-20 April 2007
Page(s):801 – 810.
[14] Mahmood, T. and Ricci, F. 2009. Improving recommender
systems with adaptive conversational strategies. In Proceedings
of the 20th ACM Conference on Hypertext and Hypermedia
(Torino, Italy, June 29 - July 01, 2009). HT '09. ACM, New
York, NY, 73-82.
[15] Sen, S., Vig, J., and Riedl, J. 2009. Tagommenders:
connecting users to items through tags. In Proceedings of the
18th international Conference on World Wide Web (Madrid,
Spain, April 20 - 24, 2009). WWW '09. ACM, New York,
NY, 671-680.
[16] Towle, B., Quinn, C. “Knowledge Based Recommender
Systems Using Explicit User Models”, In Knowledge-Based
Electronic Markets, Papers from the AAAI Workshop, Menlo
Park, CA: AAAI Press, 2000.
[17] P. Bhamidipati and K. Karlapalem, Kshitij: A Search and
Page Recommendation System for Wikipedia, COMAD 2008.
[18] Xin Sui, Suozhu Wang, Zhaowei Li, "Research on the model
of Integration with Semantic Web and Agent Personalized
Recommendation System," cscwd, pp.233-237, 2009 13th
International Conference on Computer Supported Cooperative
Work in Design, April 22-24, 2009.
[19] Yae Dai, HongWu Ye, SongJie Gong, "Personalized
Recommendation Algorithm Using User Demography
Information," wkdd, pp.100-103, 2009 Second International
Workshop on Knowledge Discovery and Data Mining, 23-25
Jan, 2009.
[20] SemMed: Applying Semantic Web to Medical
Recommendation Systems Rodriguez, A.; Jimenez, E.;
Fernandez, J.; Eccius, M.; Gomez, J.M.; Alor-Hernandez, G.;
Posada-Gomez, R.; Laufer, C., 2009. INTENSIVE '09. First
International Conference on Intensive Applications and
Services ,20-25 April 2009, Page(s):47 – 52.
[21] Richards, D., Taylor, M., and Porte, J. 2009. Practically
intelligent agents aiding human intelligence. In Proceedings of
the 8th international Conference on Autonomous Agents and
Multiagent Systems - Volume 2 (Budapest, Hungary, May 10
- 15, 2009). International Conference on Autonomous Agents.
International Foundation for Autonomous Agents and
Multiagent Systems, Richland, SC, 1161-1162.
[22] B. A. Gobin, and R. K. Subramanian,"Knowledge Modelling
for a Hotel Recommendation System", World Academy of
Science, Engineering and Technology, 2007.
[23] Chen, L. and Pu, P. 2009. Interaction design guidelines on
critiquing-based recommender systems. User Modeling and
User-Adapted Interaction 19, 3 (Aug. 2009), 167-206.
[24] Riedl. John, and Dourish. P.,”Introduction to the special
section on recommender systems“, ACM Trans. Comput.Hum. Interact. , Sep. 2005, Vol. 12 No.3, pp. 371-373.
[25] Resnick, P. and Varan, R. Hal, (Mar 1997),
Recommendation systems, Communications in ACM, Vol.40
No.3, pp 56-58.
66
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Island Model based Differential Evolution Algorithm for
Neural Network Training
Htet Thazin Tike Thein
University of Computer Studies, Yangon
Yangon, Myanmar
[email protected]
topology of the network is allowed to vary freely it can take
take the shape of any broken curve [15]. Several types of
learning algorithm have been used for neural network in the
the literature.
The DE algorithm is a heuristic algorithm for global
optimization. It was introduced several years ago (in 1997)
and has been developed intensively in recent years [16]. Its
advantages are as follows: the possibility of finding the
global minimum of a multimodal function regardless of the
initial values of its parameters, quick convergence, and the
small number of parameters that needs to be set up at the
start of the algorithm’s operation [17].
Abstract
There exist many approaches to training neural network. In this
system, training for feed forward neural network is introduced by
using island model based differential evolution. Differential
Evolution (DE) has been used to determine optimal value for
ANN parameters such as learning rate and momentum rate and
also for weight optimization. Island model used multiple
subpopulations and exchanges the individual to boost the overall
performance of the algorithm. In this paper, four programs have
developed; Island Differential Evolution Neural Network
(IDENN), Differential Evolution Neural Network (DENN),
Genetic Algorithm Neural Network (GANN) and Particle Swarm
Optimization with Neural Network (PSONN) to probe the impact
of these methods on ANN learning using various datasets. The
results have revealed that IDENN has given quite promising
results in terms of convergence rate smaller errors compared to
DENN, PSONN and GANN.
Keywords: Artificial neural network, Island Model, Differential
Evolution, Particle Swarm Optimization, Genetic Algorithm.
Differential evolution is a relatively new global search and
optimization algorithm that is suitable for the real variable
optimization. It used the vector difference and elite
selection for the selection process and have a relatively few
parameter compared to other evolutionary algorithm.
Neural network weight can be trained or optimized using
differential evolution. Island based model works by running
multiple algorithms and shares the results at regular interval
promoting the overall performance of the algorithm. This
system will propose the island based differential evolution
algorithm for training feed forward neural network.
1. Introduction
A neural network is a computing system made up of a
number of simple, interconnected processing neurons or
elements, which process information by its dynamic state
response to external inputs [1]. The development and
application of neural networks are unlimited as it spans a
wide variety of fields. This could be attributed to the fact
that these networks are attempts to model the capabilities of
of human. It had successfully implemented in the real world
world application which are accounting and finance [2,3],
[2,3], health and medicine [4,5], engineering and
manufacturing
[6,7], marketing [8,9] and general
applications [10,11,12]. Most papers concerning the use of
of neural networks have applied a multilayered, feedforward, fully connected network of perceptions [13, 14].
Reasons for the use of simple neural networks are
done by the simplicity of the theory, ease of
programming, good results and because this type of NN
represents an universal function in the sense that if the
2. Literature Review
The most widely used method of training for feed forward
ANNs is back propagation (BP) algorithm [18]. Feed
forward ANNs are commonly used for function
approximation and pattern classifications. Back
propagation algorithm and its variations such as Quick
Prop [19] and RProp [20] are likely to reach local minima
especially in case that the error surface is rugged. In
addition, the efficiency of BP methods depends on the
selection of appropriate learning parameters. The other
training methods for feed forward ANNs include those
that are based on evolutionary computation and heuristic
67
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
the evolutionary theory, PSO is based on the principles that
that flock of birds, school of fish, or swarm of bee’s
searches for food sources where at the beginning the perfect
perfect location is not known. However, they eventually
eventually they reach the best location of food source
by means of communicating with each other.
principles such as Differential Evolution (DE), Genetic
Algorithm (GA), and Particle Swarm Optimization
(PSO).
2.1 Artificial Neural Network (ANN)
An Artificial Neural Network, often just called a neural
network, is a mathematical model inspired by biological
neural networks. A neural network consists of an
interconnected group of artificial neurons, and it
processes information using a connectionist approaches to
computation. In most cases a neural network is an
adaptive system that changes its structure during a
learning phase. Neural networks are used to model
complex relationships between inputs and outputs or to
find patterns in data.
2.4 Genetic Algorithm
Genetic algorithms are stochastic search techniques that
guide a population of solutions towards an optimum using
the principles of evolution and natural genetics. In recent
years, genetic algorithms have become a popular
optimization tool for many areas of research, including
the field of system control, control design, science and
engineering. Significant research exists concerning genetic
algorithms for control design and off-line controller
analysis.
Genetic algorithms are inspired by the evolution of
populations. In a particular environment, individuals which
better fit the environment will be able to survive and hand
down their chromosomes to their descendants, while less fit
individuals will become extinct. The aim of genetic
algorithms is to use simple representations to encode
complex structures and simple operations to improve
these structures. Genetic algorithms therefore are
characterized by their representation and operators. In the
original genetic algorithm an individual chromosome is
represented by a binary string. The bits of each string are
called genes and their varying values alleles. A group of
individual chromosomes are called a population. Basic
genetic operators include reproduction, crossover and
mutation [28]. Genetic algorithms are especially capable of
handling problems in which the objective function is
discontinuous or non differentiable, non convex,
multimodal or noisy. Since the algorithms operate on a
population instead of a single point in the search space, they
climb many peaks in parallel and therefore reduce the
probability of finding local minima.
2.2 Differential Evolution (DE)
Differential evolution (DE) algorithm is a simple
evolutionary algorithm that creates new candidate
solutions by combining the parent individual and several
other individuals of the same population. A candidate
replaces the parent only if it has better fitness. This is
rather greedy selection scheme that often outperforms the
traditional evolutionary algorithm. In addition, DE is a
simple yet powerful population based, direct search
algorithm with the generation and test feature for globally
optimizing functions using real valued parameters.
Among DE’s advantages are its simple structure, ease of
use, speed and robustness. Due to these advantages, it has
many real-world applications. DE starts with random
population as like other evolutionary algorithm. Solutions
are encoded using chromosomes, for neural network
training, weight of neural network are encodes in the
chromosome. For each iteration of DE, fitness of each
chromosome is evaluated. Fitness determines the quality
of solution or chromosomes. For training neural network,
fitness function is generally MSE (mean square error) of
neural network. Each chromosome undergoes mutation
and crossover operation to produce trial vector. Each
fitness of trial vector is compared with the parent vector
and the one with greater fitness survived and the next
generations begin.
2.5 Island Model (IM)
An island model (IM) is an approach to distribute EA. It
divides individuals into subpopulations and allows for
occasional exchange of individuals (migrations). The
simplest island mode assumes the same global parameters
for islands and the same global parameters for migrations.
Populations are characterized by their number, size and
the evolutionary algorithm type. Migrations are described
by the topology.
2.3 Particle swarm optimization (PSO)
Particle swarm optimization (PSO) [21] [22] is a
stochastically global optimization method that belongs to
the family of Swarm Intelligence and Artificial Life. Similar
Similar to artificial neural network (ANN) and Genetic
Algorithms (GA) [23][24] which is the simplified models
models of the neural system & the natural selections of the
68
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
individuals are replaced by the newly obtained
individuals. Four migration strategies are common:
• Select the best individuals replace the worst
individuals.
• Select random individuals, replace the worst
individuals.
• Select the best individuals replace random individuals.
• Select random individuals, replace random
individuals.
This system experiments the best individuals replace the
worst individuals.
3. Island Model based Differential Evolution
Algorithm (IDE)
The DE algorithm was proposed by Price and Storn [25].
The DE algorithm has the following advantages over the
traditional genetic algorithm: it is easy to use and it has
efficient memory utilization, lower computational
complexity (it scales better when handling large
problems), and lower computational effort (faster
convergence) [26]. DE is quite effective in nonlinear
constraint optimization and is also useful for optimizing
multimodal problems [27].
Its pseudocode form is as follows:
3.3. Migration Interval
In order to distribute information about good individuals
among the islands, migration has to take place. This can
either be done in synchronous way every nth generation or
in an asynchronous way, meaning migration takes place
at non-periodical times. It is commonly accepted that a
more frequent migration leads to a higher selection
pressure and therefore a faster convergence. But as always
with a higher selection pressure comes the susceptibility
to get stuck in local optima. In this system, various
migration intervals will be experimented to find the best
solution for the neural network training.
a) Create an initial population consisting of
PopSize individuals
b) While (termination criterion is not satisfied)
Do Begin
c) For each ith individual in the population
Begin
d) Randomly generate three integer numbers:
r1,r2,r3∈[1;PopSize], where r1≠r2≠r3≠i
e) For each jth gene in ith individual (j∈[1;n])
Begin
vi,j =xr1,j +F•(xr2,j −xr3,j)
f) Randomly generate one real number randj∈[0; 1)
g) If rand j<CR then ui,j :=vi,j
Else ui, j: =xi, j
End;
h) If individual u i is better than individual xi then
replace individual xi by child ui individual
End;
End;
3.4. Migration Size
A further important factor is the number of individuals
which are exchanged. According to these studies the
migration size has to be adapted to the size of a
subpopulation of an island. When one migrates only a
very small percentage, the influence of the exchange is
negligible but if too much individuals are migrated, these
new individuals take over the existing population, leading
to a decrease of the global diversity. In this system,
migration size will also be investigated which can yield
the best performance.
3.1. Migration Topology
4. System Design
There are four types of migration topology. They are ring,
torus, random and fully connected topology. This system
investigates the ring topology.
3.2. Migration Strategy
A migration strategy consists of two parts. The first part is
the selection of individuals, which shall be migrated to
another island. The second part is to choose which
69
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Iris. The results for each dataset are compared and
analyzed based on the convergence rate and classification
performance. All algorithm are run for different numbers of
iteration, among of them, MSE (Mean Square Error) of
Island DE is much lower than other algorithms.
5.1 Results on XOR Dataset
Table 1: Result of IDENN, DENN, PSONN and GANN on XOR Dataset
IDENN
DENN
PSONN
GANN
Learning
Iteration
20
41
51
61
Error
Convergence
0.003
0.0048865
0.00473763
0.04125
Convergence
Time
4 sec
7 sec
12 sec
37 sec
Classification
(%)
99.2
98.97
95.17
85.66
Fig. 1 System Flow Diagram
Island model used different subpopulation with each own
island. Each island operates its own execution as like in
DE algorithm. Each island initializes the population at
the start of the algorithm or replace the subpopulation
migrates from other neighbor. Mutation, crossover and
selection are performed on the individual chromosome. If
the migration interval is not fired, the next iteration
begins within island, otherwise, a portion of its own
population and neighbor is selected for migration. If the
migration occurs, island sends sub-population to neighbor
island. Neighbor island replaces the sub-population send
by its neighbor and replace with its portion of population
and algorithm continue.
5. EXPERIMENTAL RESULT
Fig. 2 MSE on XOR dataset
Currently the system experiment the island model with
simple ring topology, migration strategy select the best
individuals replace the worst individuals. The island model
used the iteration as the migration interval and one-third of
the old population is used to migrate and replace. Learning
rate of this system is set to 0.01. Four programs have been
developed: Island Differential Evolution Feed Forward
Neural Network (IDENN), Differential Evolution Feed
Forward Neural Network (DENN), Particle Swarm
Optimization Feed Forward Neural Network (PSONN)
and Genetic Algorithm Feed Forward Neural Network
(GANN) using four dataset: XOR, Cancer, heart and
5.2 Results on Cancer Dataset
Table 2: Result of IDENN, DENN, PSONN and GANN on Cancer Dataset
Learning Iteration
Error
Convergence
IDENN
DENN
PSONN
GANN
200
443
219
10000
0.00201
0.00499
0.004870
0.50049
70
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Convergence
Time
103 sec
195 sec
110 sec
273 sec
Classification (%)
99.01
98.40
98.65
97.73
Fig. 4 MSE on Iris dataset
5.4 Results on Heart Dataset
Table 4: Result of IDENN, DENN, PSONN and GANN on Heart Dataset
IDENN
DENN
PSONN
GANN
Learning Iteration
40
58
10000
9000
Error
Convergence
0.039
0.048925
1.46392
3.00
Convergence
Time
7 sec
16 sec
170 sec
110 sec
Classification (%)
88.93
85.50
89.56
92.83
Fig. 3 MSE on Cancer dataset
5.3 Results on Iris Dataset
Table 3: Result of IDENN, DENN, PSONN and GANN on Iris Dataset
Learning
Iteration
IDENN
DENN
PSONN
GANN
28
61
818
10000
Error
Convergence
0.0205
0.049803
0.049994
1.88831
Convergence
Time
5 sec
16 sec
170 sec
256sec
Classification
(%)
96.39
95.014972
93.86
97.72
Fig. 5 MSE on Heart dataset
71
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
6. Comparison IDENN, DENN, PSONN and
GANN
References
[1] A. Majdi, M. Beiki: Evolving Neural Network Using Genetic
Algorithm for Predicting the Deformation Modulus of Rock
Masses. International Journal of Rock Mechanics and Mining
Science, vol. 47(2), pp. 246-253, 2010.
[2] Lee, K., Booth, D., and Alam, P.: A Comparison of Supervised
and Unsupervised Neural Networks in Predicting Bankruptcy of
Korean Firms. Expert Systems with Applications, vol. 29(1), pp.1–
16, 2005.
[3] Landajo, M., Andres, J. D., and Lorca, P.: Robust Neural
Modeling for the Cross-Sectional Analysis of Accounting
Information. European Journal of Operational Research, vol.
177(2), pp. 1232–1252, 2007.
[4] Razi, M. A., and Athappily, K.: A Comparative Predictive
Analysis of Neural Networks (NNs), Nonlinear Regression and
Classification and Regression Tree (CART) Models. Expert
Systems with Applications, vol. 2(1), pp. 65–74, 2005.
[5] Behrman, M., Linder, R., Assadi, A. H., Stacey, B. R., and
Backonja, M. M.: Classification of Patients with Pain Based on
Neuropathic Pain Symptoms: Comparison of an Artificial Neural
Network against an Established Scoring System. European Journal
of Pain, vol. 11(4), pp. 370–376, 2007.
[6] Yesilnacar, E., and Topal, T.: Landslide Susceptibility
Mapping: A Comparison of Logistic Regression and Neural
Networks Methods in a Medium Scale Study, Hendek region
(Turkey). Engineering Geology, vol. 79 (3–4), pp. 251–266,
2005.
[7] Dvir, D., Ben-Davidb, A., Sadehb, A., and Shenhar, A. J.:
Critical Managerial Factors Affecting Defense Projects Success: A
Comparison between Neural Network and Regression Analysis.
Engineering Applications of Artificial Intelligence, vol. 19, pp.
535–543, 2006.
[8] Gan, C., Limsombunchai, V., Clemes, M., and Weng, A.:
Consumer Choice Prediction: Artificial Neural Networks Versus
Logistic Models. Journal of Social Sciences, vol. 1(4), pp. 211–
219, 2005.
[9] Chiang, W. K., Zhang, D., and Zhou, L.: Predicting and
Explaining Patronage Behavior toward Web and Traditional Stores
Using Neural Networks: A Comparative Analysis with Logistic
Regression. Decision Support Systems, vol. 41, pp. 514–531,
2006.
[10] Chang, L. Y.: Analysis of Freeway Accident Frequencies:
Negative Binomial Regression versus Artificial Neural Network.
Safety Science, vol. 43, pp. 541–557, 2005.
[11] Sharda, R., and Delen, D.: Predicting box-office success of
motion pictures with neural networks. Expert Systems with
Applications, vol. 30, pp. 243–254 , 2006.
[12] Nikolopoulos, K., Goodwin, P., Patelis, A., and
ssimakopoulos, V.: Forecasting with cue information: A
comparison of multiple regressions with alternative forecasting
approaches. European Journal of Operational Research, vol.
180(1), pp. 354–368, 2007.
[13] S. Curteanu and M. Cazacu.: Neural Networks and Genetic
Algorithms Used For Modeling and Optimization of the Siloxane Siloxane Copolymers Synthesis. Journal of Macromolecular
Science, Part A, vol. 45, pp. 123–136, 2007.
[14] C. Lisa and S. Curteanu: Neural Network Based Predictions
for the Liquid Crystal Properties of Organic Compounds.
Fig. 6 Comparison of correct classification Percentage IDENN, DENN,
PSONN and GANN
For XOR dataset, the results show that IDENN has better
results on convergence time and correct classification
percentage. IDENN convergence in a short time with high
correct classification percentage. For Cancer dataset,
IDENN classification results are better than DENN,
PSONN and GANN. For Iris dataset, GANN classification
results are better than IDENN, DENN, and PSONN. For
Heart dataset, GANN classification results are better than
IDENN, DENN, and PSONN. For overall performance,
the experiments show that IDENN significantly reduces the
error with minimum iterations. IDENN produces feasible
results in terms of convergence time and classification
percentage.
7. Conclusion
This system presents the neural network training
algorithm using island model based differential
algorithm. By exploiting the global search power of
differential evolution algorithm in conjunction with island
model will boost the training performance of the
algorithm. The system will converge quickly to the lower
mean square error. Island model encourage the diversity
among the individual among islands which increase
search capability and by migration island model can share
the best experiences of each other. By using island model
rather than single DE, it can get advantages from parallel
problem solving and information sharing which lead to
faster global search.
72
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Computer-Aided Chemical Engineering, vol. 24, pp. 39–45,
2007.
[15] Fernandes and L.M.F. Lona: Neural Network Applications in
Polymerization
Processes, Brazilian Journal Chemical
Engineering, vol. 22, pp. 323–330, 2005.
[16] K. V. Price, R. M. Storn, and J. A. Lampinen, Differential
Evolution: APractical Approach to Global Optimization. Berlin,
Germany: Springer Verlag, 2005.
[17] R. L. Becerra and C. A. Coello Coello, “Cultured differential
evolution for constrained optimization,” Comput. Methods Appl.
Mech. Eng., vol. 195,no. 33–36, pp. 4303–4322, Jul. 1, 2006.
[18] Hecht Nelso R., back propagation neural network, 1989.
[19] Scott E. Fahlman: An Empirical Study of Learning Speed in
Back-Propagation Networks, September 1988.
[20] Rprop – Description and Implementation Details Martin
Riedmiller, Technical report, 1994.
[21] Kennedy, J.; Eberhart, R. "Particle Swarm Optimization".
Proceedings of IEEE International Conference on Neural
Networks, 1995.
[22] Kennedy, J.; Eberhart, R.C. Swarm Intelligence. Morgan
Kaufmann, 2001.
[23] Zhang, G.P., Neural networks for classification: a survey.
IEEE Transactions on Systems Man and Cybernetics, 2000.
[24] Rudolph, G., Local convergence rates of simple evolutionary
algorithms with cauchy mutations, 1997.
[25] R. Storn and K. Price, “Differential evolution—A simple and
efficient heuristic for global optimization over continuous spaces,”J.
Glob. Optim. vol. 11, no. 4, pp. 341–359, Dec. 1997.
[26] K. V. Price, R. M. Storn, and J. A.Lampinen, Differential
Evolution—A Practical Approach to Global Optimization. Berlin,
Germany: SpringerVerlag, 2005.
[27] D. Karaboga and S. Okdem, “A simple and global
optimization algorithm for engineering problems: Differential
evolution algorithm,” Turkish J. Elect. Eng. Comput. Sci., vol. 12,
no. 1, pp. 53–60, 2004.
[28] D. Whitley, "Applying Genetic Algorithms to Neural Network
Problems," International Neural Network Society p. 230, 1988.
Htet Thazin Tike Thein received her B.C.Tech. degree in Computer
Technology from University of Computer Studies in Yangon Myanmar
and M.C.Tech. degree in Computer Technology from University of
Computer Studies in Yangon Myanmar. She is currently a student of
Ph.D. degree at University of Computer Studies in Yangon Myanmar.
She is currently been working toward her Ph.D. degree in Neural
Network and classification.
73
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
3D-Diagram for Managing Industrial Sites Information
Hamoon Fathi1, Aras Fathi2
1
Stractural engineering Department of Kurdistan University- ASCE member
Sanandaj, IRAN
[email protected]
2
Food Technology engineering Department of Islamic Azad University
Science and research Branch, Tehran, IRAN
[email protected]
Abstract
2. Methodology
This paper proposes a 3D-model for managing industrial sites
informations with coding and prioritizing the sources. Managing
the industrial and construction site information are organized by
PMBOK standard in this study. This model investigated on three
different construction sites and factories including a building site,
a tractor assembling factory and a food production factory. The
results show, the 3D-model has floating act in the different
problems. The proposed model is in good agreement with site
information results.
Keywords: 3D-modeling, industrial site, information
management, PMBOK standard.
This study proposes a 3D-model for managing the
industrial site information. Methodology of the model is
focused on PMBOK standards. These methodologies
combine the sources by coding and prioritizing. Source
types in different projects are categorized with same
standards. Cost, timetable, communication, risk and
limitations are some parts of the uniform management at
PMBOK standards. Human sources and quality are
prioritized with the other sources at the same level (Fig1).
1. Introduction
3D-diagram is one of the new automate methods for
managing the industrial information and risk management.
Many models used for managing the construction of
projects are based on systematic structure [1-5]. Some
researches tried to develop experience and knowledge of
modeling the construction management [6-8]. The main
purpose of 3D-diagram is managing and estimating the
risks of the project. However, this information can build a
way for analyzing the future of the project. Many
researchers studied the risk of construction operation
analysis [9-18]. Some of these studies discussed about
industries risk and cost management [18-25]. Reduce cost
and time and improve the quality are object of engineering
simulations. Site information and source data are the feed
of engineering simulations. This study assemblage the site
information and source data in the 3D-diagrams. This
information is gathered from the industrial sites. Several
researches tried to simulate the site information with
computer systems [26-31]. Computer software and models
help to simulate the project dynamically. Large scale
projects belong to classify the complex dynamic systems.
Traditionally, formal modeling tools are not dynamic so
the 3D-models are used to improve and build dynamic
management. Furthermore, producing schedules for a
project by computer needs a uniform format for translating
the information to the managers. 3D-diagrams' algorithm
supply the uniform format for managing site information.
This study proposes a 3D-diagrams' algorithm for risk,
cost and time management. Three different industrial
projects tested by 3D-diagrams for different kinds of
source management. The manager limitations defined with
PMBOK standards.
Fig. 1 Uniform management at PMBOK standards
2.1 Cost management
Budget sources are one of the important parts of the
projects and the other sources have direct relation with
cost. Cost management needs estimating, predicating and
controlling all duration of the project's time. Cost is one of
the dimensions at this methodology and growing with
uniform gradient. Estimating techniques for cost of the
activities need to combine all the sources and contractor's
responsibilities for their works however normal project
cost would be predictable by suitable sub programs.
2.2 Time management
Most of the projects are programed with time, whereas
managing the project time has relationship with cost. Most
projects have a simple diagram for estimating the costs or
incomes. However, all the sources' information is used for
simulating cost-time diagram. The other sources are
installed by coding and prioritizing in the third direction of
this methodology. In fact, the normal cost-time diagram is
74
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Normally, value of weight-cost items is important in
critical and risk points in project management. Items
which have some parts under the budget cover in the 3Ddiagram are financed by the other sources. Items which are
completely out of the budget line are contract and parttime items. It should be considered that the budget of these
items depends to the contract’s type. Many design
software are used for estimate the project behavior.
However, most of these programs are based on time and
lost budget. The 3D-model algorithm can be used by
mathematical software beside the other managing project
softwares.
unqualified for balancing between time and budget alone.
Because, the cost behaviors are usually not linear over the
full range of activities and the cost go up in the inactive
parts.
3. Algorithm of the 3D-model
Most of the uniform management stabilized on the
timetable. Time is the determination factor for calculating
project improvement in 2D-modeling. Controlling by
timetable has some limitations such as: time doesn’t have
negative form and growing without any activities. Human
management and quality are two activities which have
cost, even while the project has stopped, so the total cost
can be another factor for calculating project improvement.
3.2 3D-models
Perfect project has activities with lag and lead relationship
which they categories in the same level with different
codes. The main algorithm is not including the temporary
and limit works, subcontracts and accidents, so the 3Dmodel forecasts these items on the costs vector
automatically. Actually, the 3D-model shows balance and
optimizes the ways between the budget and cost points of
the project at different times. Figure 5, provides a colored
3D-model of a sample project. Risk management and
managing the critical conditions are based on cost and
time perform by changing the vector of the project from
start point to end point. In fact, the changing of the project
vector is used for decreasing the time or cost of the
project.
3D-model applies the both factors of cost and time for
managing the industrial site informations. The proposed
model estimates the project’s behavior with using
combinations of cost and time. PMBOK standards present
the usual curve between total cost and time in normal
constructions (Fig 2). Total cost and time in the project
have relationship, and this relation is depended on the
other activities and sources.
These reductions participate with erasing some parts of the
activities or growing the cost or time. One of the most
important abilities of the 3D-model management is
estimating the effect of cost on the project duration. For
example, Figure 5 shows the sample project with recession
on the cost sources. These methods are used by
commander and senior of the project. The sector managers
have a WBS for their own activities and input them in the
special diagrams.
Fig. 2 Cost –Time curve
3.1 Algorithm
The algorithm of the 3D-model of management site is
simple and each activity divided in the branch with
specified height and those items are input in the algorithm
with priority (Fig 3). Each activity has a millstone and
work break structure (WBS). Sub management of the
WBS is characterized with codes for the items. Whole way
of the project started from the meet point of cost and time
vectors and ended at the opposite point at the total cost and
time (Fig 4).
3D-diagram is a profile with three dimensions. Cost is
arranged by one of the axis and time by the other one. The
third axis pertains to the other sources. A project’s
approach begins from the origin of the coordinate timecost curve to the end point. Direct platform from the start
point to the end of the project constitute an S-form shape
which have been defined by PMBOK standards. This
approach conducts the managers to the shortest vector.
Managers shift the project’s approach to control cost or
time.
Fig. 4 The algorithm of WBS (subprogram)
75
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Fig. 3 The algorithm of the management site's 3D-model
Fig. 5 3D-model of a sample project
76
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
This part investigated a tractor assembling factory at
Sanandaj city with 3D-modeling management (Fig 8).
The task of this factory is assembling and repairing their
products. One of the important parts at the assembling
factory is the human source management. Table 2 is
including the work numbers, wages and abilities and
some input information for 3D-diagram. The investigation
results with 3D-modeling of the human sources
management are shown (Fig 9). Daily workers' numbers,
numbers of skilled workers and engineers, amount of
overtime hours, upload and unload the employees are
characterized in figure 9. This 3D-diagram categorizes the
employees to nine groups. These groups are characterized
with E1 to E9 symbols. Numbers of working days, row
fiscal and start to end working periods are specified in the
3D-diagram by the means of numbers and percentages.
Negative percentages replaced with zero number and
positive percentages are cumulatively shown in the last
step of the chart. This 3D-diagram submits the minimum
to maximum salary increment for E1 to E9 groups while
the E9 has the maximum payment. In fact, E9 is showing
the manager team.
4. Test Results and Discussion
These parts are discussing about 3D-modeling results and
applications, on the real industrial site information. Data
and information are used to develop the 3D-model in the
third direction and the other directions are cost and time.
The purpose of this model is to provide comprehensive
guidance for information's bank on risk management.
4.1 building construction site
A residential tower with 80 houses is controlled by the
cost-time 3D-model. These houses have the same
infrastructures and been built in an apartment with 12
floors at Sanandaj city (Fig6). The project started at May
2009 and finished at March of 2013. Supply and
producing materials in the building site is one of the
important items. Table 1 is including some parts of the
required materials, cost and budget, sources and storage
capacity in the limited site at the center of a city (input
data). The site environment had effect on the time and
cost of the budget. Figure 7, presents the time of
producing, type and value, sources and daily cost of the
materials and some other information in the 3D-diagram.
The 3D-diagram is including time, cost and materials.
Materials of the construction project were classified into
16 groups on an axis. M1 to M5 groups are non-storable
material and used upon receipt to the construction site.
The E6 is related to night time shift workers with 70 %
bonus adjust the salary base. Risk management is one of
this diagrams consequences. The amount of personnel
costs arrived to the critical point at autumn and winter of
2011, so one of the solutions could be reducing some
parts of night time shifts in the mentioned seasons to
decrease the total cost from the critical limitations by the
mangers.
M6 to M16 are storable material which assort according
to use time and storage’s maximum capacity. This
diagram contains a large group of information, for
example M14 group’s material, are loaded in a constant
specified amount from the project’s start point (May
2009) to the end line (March 2013). Therefore the M14’s
unload and reload is a constant number. M14 determines
water in this project. This project is a sample and some of
the management results are:
 M15 and M16 material, are symbols for doors,
valves, pipes and decoration’s material that been
logged in the last year of the project duration. As
the storage capacity was not considered to them,
so they temporary stored in some of the
building’s rooms. Accordingly, material
purchasing postponed to construction of the
building roofs, annual inflection causes
increasing cost in material providing.
 Most parts of entering materials are stored
discontinuously and new purchasing will be done
after the most part of the material been used
approximately. This process is able the
managing team to assign a large part of the
storehouse to expensive and importing material.
However depletion of unsorted material could be
counted as a risk and cause an increase in the
project’s duration.
Cumulative part of the salaries shows equality between
the total and critical cost. The maximum value is due to
repairing the tractor products on the assembly line.
Actually, environmental condition and temperature affects
the workers efficiency. For this reason, the factory needs
to postpone the personnel adjustment or funding for staff
salaries in previous months. One of the time-cost 3Ddiagram’s functionalities is to define the risks and finding
best solutions.
4.3 Food production factory
Cost and quality are dependent parameters which they
have direct effect on the find product and marketing (Fig
10). One of the most important parts of the food factories
is quality control. Managing the quality attend to
marketing results and healthy controls. Figure 11,
suggests the optimum cost of investigation at the different
properties of the final food product. This kind of 3Ddiagram is mostly used by sales and marketing senior
executives. These executives in cooperating with R&D,
decide on the size, flavors, advertising and product titles.
Results of this diagram, presents a long-term program for
sale’s improvement. Of course, these results are not
available until the end of one complete sale period.
Product’s sale duration, total costs, consumer’s feedback
and other sale information, are forming the 3D-diagram.
4.2 Tractor assembling factory
77
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Table 1: Materials information and input data
Material
Name
Start point
End point
Budget
cost
Capacity
Value
Concrete
welding
Plasticizer
Casing
Installation
Stucco
Brickwork
wiring
Tarry
Alabaster
M1
M2
M3
M4
M5
M6
M7
M8
M9
M10
2009-10
2009-10
2009-10
2009-5
2010-1
2009-5
2009-10
2011-3
2010-9
2010-1
2011-10
2011-7
2011-7
2011-10
2011-1
2012-12
2012-10
2012-8
2012-6
2012-5
59000
15000
24600
24000
12000
21000
18000
8500
8500
7000
3000
3000
3000
3000
3000
1000
1000
1000
1000
1000
Non-storable
Non-storable
Non-storable
Non-storable
Non-storable
8 parts
7 parts
6 parts
6 parts
6 parts
500
0
1100
0
0
1000
1000
700
700
500
Scoria
Calk
Crockery
Water
Window
Door
M11
M12
M13
M14
M15
M16
2010-1
2010-1
2011-4
2009-5
2012-10
2012-10
2012-5
2012-2
2013-1
2013-3
2013-3
2013-3
7000
5000
8500
10000
13000
13000
1000
1000
1000
1000
1000
1000
6 parts
5 parts
5 parts
10 parts
4 parts
4 parts
500
0
500
0
1500
1500
Fig. 6 A residential tower with 80 houses
Table 2: Employee's information and input data
Group
Activity
Number
E1
E2
E3
E4
E5
E6
E7
E8
E9
Guardsman
Journeyman
Serviceman
Repairman
Engineer
Night Workers
Engineer
Inspector
Manager
3
12
5
5
7
7
5
6
3
Working
day
455
455
225
180
360
455
270
225
180
Wage
Start point
End point
7
14
21
21
30
32
37
47
60
2011-1-1
2011-1-1
2011-2-15
2011-6-15
2011-3-1
2011-1-1
2011-3-10
2011-3-1
2011-7-1
2012-3-28
2012-3-28
2011-7-15
2011-9-28
2012-1-15
2012-3-28
2011-10-1
2012-1-28
2012-3-28
Fig. 8 Tractor assembling factory
78
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Fig. 7 The 3D-diagram of time, cost and materials
Fig. 9 3D-modeling of the human sources management
79
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Fig. 11 3D-diagram of food production factory
Fig. 10 food production factory
For example, this diagram includes four items that each
one has two equilibrium coefficients to balance the price
and numbers. Presented coefficient proposes information
to balance the items at a uniform mathematical form. Size
and shape of the products is one of these marketing items.
Results describe that alternative size does not have any
significant effect on the sale rate firstly, however coincide
sale of six different sizes together caused growing sale
rate.
5. Conclusions
This study, investigated the feasibility of using 3Ddiagrams for collecting the information of site project in a
uniform format. The methodology of target reduces time
and cost in the risk management. The results focused on
real industrial information to analyze project conditions
before and after constrictions. The following conclusions
can be drawn from the experimental results and new
model:
 The 3D-diagrams for risk management are
rational and compatible with industrial results
and informations.
 The 3D-diagrams analyses the project conditions
with cost and time at the same level of valuation,
so the diagrams estimate the future to manage
the project and rid deleterious items.
Helping to decide between repeats and diversity of
advertisements is one of this diagrams consequences.
Actually advertising is the main factor to recommend the
product to consumers. Investigations show that efficiency
of an advertisement decreases after a repeating period.
The suitable time to change the advertisement could be
taken from the 3D-diagram. In this case after 72 repeats
of an advertisement, the sale’s growing procedure stops
denoting that more repeating could not be commodious
anymore.
80
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
[17] H. Tanaka, L.T. Fan, F.S. Lai and K. Toguchi, "Fault tree
analysis by fuzzy probability", IEEE Transactions on
Reliability, Vol. 32, No. 5, 1983, pp. 453-457.
[18] S. Azampanah, A. Eskandari, A. Khademzadeh and F.
Karimi, "Traffic-Aware Selection Strategy for ApplicationSpecific 3D NoC", Advances in Computer Science: an
International Journal, Vol. 2, Issue 4, No.5, September 2013,
pp. 107-114.
[19] L.A. Zadeh, "Probability of measure of fuzzy events",
Journal of Mathematical Analysis and Application, Vol. 23,
1968, pp. 421-427.
[20] M.I. Kellner, R.J. Madachy and D.M. Raffo, "Software
process simulation modeling: Why? What? How?" Journal
of Systems and Software, 46(2-3), 1999, April 15, pp. 91105.
[21] G. Rincon, M. Alvarez, M. Perez and S. Hernandez, "A
discrete-event simulation and continuous software
evaluation on a systemic quality model: An oil industry
case", Information & Management, 42, 2005, pp. 10511066.
[22] J.K. Shank and V. Govindarajan, "Strategic Cost
Management: The Value Chain Perspective", Journal of
Management Accounting Research 4, 1992, pp. 179-197.
[23] A. Maltz and L. Ellram, "Total cost of relationship: An
analytical framework for the logistics outsourcing decision",
Journal of Business Logistics 18 (1), 1997, pp. 45-55
[24] L.M. Ellram, "A Framework for Total Cost of Ownership",
International Journal of Logistics Management 4 (2), 1993,
pp. 49-60.
[25] B. Boehm and R. Turner, "Using Risk to Balance Agile and
Plan-Driven Methods", IEEE Computer 36(6), June 2003,
pp. 57-66.
[26] K.M. Chandy and J. Misra, "Distributed Simulation: A
Case Study in Design and Verification of Distributed
Programs", IEEE Transactions on Software Engineering SE5(5), 1978, pp. 440-452.
[27] R.M. Fujimoto, "Performance Measurements of Distributed
Simulation Strategies", Transactions of the Society for
Computer Simulation 6(2), 1989, pp. 89-132.
[28] H. Mehl, M. Abrams and P. Reynolds, "A Deterministic
Tie-Breaking Scheme for Sequential and Distributed
Simulation: Proceedings of the Workshop on Parallel and
Distributed Simulation", Journal Society for Computer
Simulation. 24, 1992, pp. 199-200.
[29] B.D. Lubachevsky, "Efficient Distributed Event-Driven
Simulations of Multiple-Loop Networks", Communications
of the ACM 32(1), 1989, pp. 111-123.
[30] P. B. Vasconcelos, "Economic growth models: symbolic
and numerical computations", Advances in Computer
Science: an International Journal, Vol. 2, Issue 5, No.6 ,
November 2013, pp. 47-54.
[31] U. Dahinden, C. Querol, J. Jaeger and M. Nilsson,
"Exploring the use of computer models in participatory
integrated assessment – experiences and recommendations
for further steps", Integrated Assessment 1, 2000, pp. 253266.
Reference
[1] P. Carrillo, "Managing knowledge: lessons from the oil and
gas sector", Construction Management and Economics, Vol.
22, No 6, 2004, pp. 631-642.
[2] I. Nonoka and H. Takeuchi, "The Knowledge Creating
Company: How Japanese Companies Create the Dynamics
of Innovation. New York", NY: Oxford University Press,
1995. 304 p.
[3] T. Maqsood, A. Finegan and D. Walker, " Applying project
histories and project learning through knowledge
management in an Australian construction company", The
Learning Organization, Vol. 13, No1, 2006, pp. 80-95.
[4] N. Bahra, "Competitive Knowledge Management. New
York", NY:Palgrave, 2001. 258 p.
[5] J.M. Kamara, A.J. Chimay and P.M. Carillo, "A CLEVER
approach to selecting a knowledge management strategy",
International Journal of Project Management, Vol. 20, No 3,
2002, pp. 205-211.
[6] A. Kaklauskas, E.K Zavadskas and L. Gargasaitė, "Expert
and Knowledge Systems and Data-Bases of the Best
Practice", Technological and Economic Development of
Economy, Vol. X, No3, 2004, pp. 88-95.
[7] K.B. DeTienne and R.B. Jensen, "Intranets and business
model innovation: managing knowledge in the virtual
organization. In: Knowledge Management and Business
Model Innovation / Y. Malhotra (Ed.). Hershey", PA: Idea
Group Publishing, 2001, pp. 198-215.
[8] S. Ogunlana, Z. Siddiqui, S. Yisa and P. Olomolaiye,
"Factors and procedures used in matching project managers
to construction projects in Bangkok", International Journal
of Project Management, Vol. 20, 2002, pp. 385-400.
[9] B.M Ayyub and A. Haldar, "Decision in construction
operation", Journal of Construction Engineering and
Management, ASCE, Vol. 111, No. 4,1985, pp. 343-357.
[10] H.N. Cho, H.H. Choi and Y.B. Kim, "A risk assessment
methodology for incorporating uncertainties using fuzzy
concepts", Reliability Engineering and System Safety, Vol.
78, 2002, pp. 173-183.
[11] H.H. Choi, H.N Cho and J.W Seo, "Risk assessment
methodology for underground construction projects",
Journal of Construction Engineering and Management,
ASCE, Vol. 130, No. 2, 2004, pp. 258-272.
[12] S. Lee and D.W. Halpin, 'Predictive tool for estimating
accident risk", Journal of Construction Engineering and
Management, ASCE, Vol. 129, No. 4, 2003, pp. 431-436.
[13] D. Singer, "A fuzzy set approach to fault tree and reliability
analysis", Fuzzy Sets and Systems, Vol. 34, 1990, pp. 145155.
[14] P.V. Suresh, A.K. Babar and V.V. Raj, "Uncertainty in
fault tree analysis: A fuzzy approach", Fuzzy Sets and
Systems, Vol. 83, 1996, pp. 135-141.
[15] J.H.M. Tah and V. Carr, "A proposal for construction
project risk assessment using fuzzy logic", Construction
Management and Economics, Vol. 18, 2000, pp. 491-500.
[16] J.H. Tah and V. Carr, "Knowledge-based approach to
construction project risk management", Journal of
Computing in Civil Engineering, ASCE, Vol. 15, No. 3,
2001, pp. 170-177.
81
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Benefits, Weaknesses, Opportunities and Risks of SaaS
adoption from Iranian organizations perspective
Tahereh Rostami1, Mohammad Kazem Akbari 2 and Morteza Sargolzai Javan 3
1
Taali University of Technology
Computer Engineering and IT Department
Qom, Iran
[email protected]
2
Amirkabir University of Technology
Computer Engineering and IT Department
Tehran, Iran
[email protected]
3
Amirkabir University of Technology
Computer Engineering and IT Department
Tehran, Iran
[email protected]
Standards and Technology (NIST), major
characteristics of cloud services are: on-demand selfservice, ubiquitous network access, location
independent resource pooling, rapid elasticity, and
measured service. Cloud services based on cloud
computing can free an organization from the burden
of having to develop and maintain large-scale IT
systems; therefore, the organization can focus on its
core business processes and implement the
supporting applications to deliver the competitive
advantages [1]. Today, cloud services have been
regarded not only as the favorable solutions to
improve an organization’s performance and
competitiveness, but also as the new business models
for providing novel ways of delivering and applying
computing services through IT. Generally, cloud
services can be divided into three categories:
Software as a Service (SaaS), Platform as a Service
(PaaS), and Infrastructure as a Service (IaaS). Among
them, SaaS is regarded as a potential segment and the
utilization of SaaS solutions can lead to many
benefits for enterprise users with profound
consequences in improving IT performance [2]. SaaS
delivers applications’ functionality through the media
of the Internet as a service [3]. Although many
vendors announced that the SaaS adoption can bring
out promising benefits, yet some organizations are
still reluctant to introduce SaaS solutions due mainly
to the trust concern (e.g., data security, network
security). In fact, each service model (SaaS, PaaS, or
IaaS) has its own security issues and calls for a
different level of security requirement in the cloud
Abstract
Software as a Service (SaaS) is a new mode of software
deployment whereby a provider licenses an application to
customers for use as a service on demand. SaaS is regarded
as a favorable solution to enhance a modern organization’s
IT performance and competitiveness which helps
organizations avoid capital expenditure and pay for the
functionality as an operational expenditure. SaaS has
received considerable attention in recent years, and an
increasing number of countries have consequently
promoted the SaaS market. However, many organizations
may still be reluctant to introduce SaaS solutions mainly
because of the trust concern they may perceive more risks
than benefits. This paper focuses on the analysis of Iranian
organizations understand from the benefits, weaknesses,
opportunities and risks of SaaS adaption.
Keywords: Cloud services, Software as a Service (SaaS),
Adoption, Binomial test
1. Introduction
Effectively making use of Information technology
(IT) can constitute a sustainable source of an
organization’s competitiveness. Cloud computing has
become a topic of tremendous interest as
organizations struggle to improve their IT
performance. Cloud services can be viewed as a
cluster of service solutions based on cloud
computing, which involves making computing, data
storage, and software services available via the
Internet. According to the U.S. National Institute of
82
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
environment [2, 4]. Some surveys related to cloud
services have enhanced our understandings of the
factors involved in adoption of SaaS solutions. For
example, in The Adoption of Software as a Service in
Small and Medium-Sized Businesses (IDC #205798,
2007), the report remarked that while SaaS has strong
growth potential, small and medium-sized businesses
have not been adopting SaaS as quickly as originally
anticipated. Concern about data security is the factor
most frequently cited as discouraging the use of
SaaS. This report also revealed that marketing efforts
for the SaaS adoption should highlight the issue of
trust by enhancing users’ perceived benefits as well
as decreasing users’ perceived risks. The rest of this
paper is organized as follows. In Section 2, presents
the related work. In Section 3, the SaaS concept has
been described. In Section 4, Effective factors on
adoption of cloud is detected. In Section 5, the
methodology has been defined.
3. Software as a Service
SaaS is an outsourcing innovation that transforms IT
resources into continuously provided services
[5].That is, SaaS delivers an application’s
functionality through the Internet as a service and
thus, eliminates the need to install and run the
software on the client’s computer [3,6]. Therefore,
customers only pay for their use of the software
because there are no licensing fees [7, 8]. This unique
feature of SaaS has allowed the SaaS market to grow
six times faster than the packaged software market
and is expected to facilitate further development of
SaaS. According to a study by Gartner, SaaS is
predicted to become increasingly important in most
enterprise application software (EAS) markets. The
global SaaS market is expected to reach 12.1 billion
USD by 2014, reflecting a compound annual growth
rate of 26%. This rapid growth of the SaaS market
has had considerable influence on the software
market [9]. However, despite this rapid growth of the
SaaS market, some of countries with SaaS markets in
their initial stages have faced many problems in SaaS
adoption. According to [9], in a new SaaS market,
inducing SaaS adoption is likely to be difficult due to
major inhibitors, such as limited integration and
flexibility.
In fact, not everyone is positive about SaaS adoption.
Some companies and market researchers are
particularly skeptical about its viability and
applicability in strong EAS markets such as ERP.
The main adoption barriers are said to be reliability
issues (i.e., stable access to services), information
security and privacy concerns (i.e., security breaches
and improper protection of firm data), and process
dependence (i.e., performance measurement and
service quality) [10]. Also there are vulnerabilities in
the applications and systems availability may lead to
the loss of valuable information and sensitive data or
may be the money. These concerns discourage the
enterprises to adopt the SaaS applications in the
cloud.
2. Related Works
In [16], presumed that SaaS adoption is a trust issue
involving perceived benefits such as pay only for
what you use, monthly payments, costs and perceived
risks such as data locality and security, network and
web application security. The paper has proposed a
solution framework that employs a modified
DEMATEL approach to cluster a number of criteria
(perceived benefits and perceived risks) into a cause
group and an effect group, respectively. In [17],
attempts to develop an explorative model that
examines important factors affecting SaaS adoption,
in order to facilitate understanding with regard to
adoption of SaaS solutions. An explorative model
using partial least squares (PLS) path modeling is
proposed and a number of hypotheses are tested,
which integrate TAM related theories with additional
imperative constructs such as marketing effort,
security and trust.in [12], analyzes the opportunities
such as Cost advantages, Strategic flexibility, focus
on core competencies and risks such as Performance,
Economic, Managerial, associated with adopting
SaaS as perceived by IT executives at adopter and
non-adopter firms. Also developed a research model
grounded in an opportunity-risk framework.
In [18], reports on research into SaaS readiness and
adoption in South Africa as an emerging economy.
Also discussed are benefits of Immediacy, Superior
IT Infrastructure, Software Maintenance and
Challenges of limited customization, integration
Problems, Perceived Security concerns.
4. Benefits, Weaknesses, Opportunities
and risks of SaaS adoption
Understand the benefits, Weaknesses, Opportunities
and risk of SaaS as a subjective manner that members
collectively assess their cloud adoption. Hence they
were identified as follows:
83
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
completely shifting responsibility for developing,
testing, and maintaining the outsourced software
application and the underlying infrastructure to the
vendor [12].
Access to specialized resources: SaaS clients benefit
from economies of skills by leveraging the skills,
resources, and capabilities that the service provider
offers. These specialized capabilities (e.g., access to
the latest technologies and IT related
Know-how) could not be generated internally if the
application were delivered in-house via an onpremises model [12].
4.1 SaaS Benefits
Access Anywhere: one of the advantages of the SaaS
is Applications used over the network are accessible
anywhere and anytime, typically with a browser.
Zero IT infrastructure: When delivering business
applications via SaaS, the complexity of the
underlying IT infrastructure is all handled SaaS
vendor.
Software Maintenance: the SaaS vendors are
responsible for any software updates; and these
happen almost without the customer noticing.
Lower cost: the cost of using SaaS can be
significantly lower compared with on premise
software, because the clients only pays for what they
use.
4.4 SaaS Risks
Lack of control risk: When the SaaS goes down,
business managers can find themselves feeling
completely helpless because they suddenly have no
visibility of the infrastructure.
Legal issues: There are legal risk include
Governance, SLAs, service reliability and
availability, etc.
Security risks: data protection and privacy are
important concerns for nearly every organization.
Hosting data under another organization’s control is
always a critical issue which requires stringent
security policies employed by Cloud providers. For
instance, financial organizations generally require
compliance with regulations involving data integrity
4.2 SaaS Weaknesses
Immature SaaS: Some feel that the SaaS model is
still immature and has yet to prove itself worthy, and
are waiting for it all to settle down before moving
forwards even if their own infrastructure is far from
perfect.
Ambiguous and complex pricing: Often providers
offer different rates for their services. Usage costs,
support and maintenance can be different. There is no
public standard tariff that all providers are required to
follow it. So consumers are confused.
Dependence on the SaaS provide: The customer is
dependent on the service provider. The service will
develop or end based on service provider’s actions.
Also If the SaaS provider were to go bankrupt and
stopped providing services, the customer could
experience problems in accessing data and therefore
potentially in business continuity.
Dependence on the Internet: In most cases, the
service cannot be used offline. The service is
available only over the Internet.
and privacy. Security and Privacy is multidimensional in nature and includes many attributes
such as protecting confidentiality and privacy, data
integrity and availability [15].
Performance risks: Performance risks are the
possibility that SaaS may not deliver the expected
level of service. Service Down time or slow it can be
a huge economic losses inflicted to the organization.
Economic risks: if the client wants to customize the
application core, he needs to own it. Even if the client
can use the standard core, he may want to build
components on top of the core functionality (using
APIs) to suit his needs with regard to integration and
customization. Higher-than-expected costs may thus
arise from the additional or changing future
requirements. In addition, increasing costs may
emerge from the hold-up, because vendor ownership
of the application core provides the vendor with more
future bargaining power. This power enables him to
increase prices, charge extra costs, or refuse to invest
in backward-compatible interfaces for the client's
customized code.
Internet resilience and bandwidth: SaaS does not
provide application availability and/or network
bandwidth as the provider originally stipulated [14].
System outages and connectivity problems can affect
4.3 SaaS Opportunities
Cost saving: No purchase of software licenses,
reduce staff IT, eliminating the cost of deployment
and infrastructure leading to savings in the overall
cost of organization.
Strategic flexibility: SaaS adoption provides a great
degree of flexibility regarding the utilization of easily
scalable IT resources. This flexibility makes it easier
for firms to respond to business-level volatility,
because the SaaS provider handles fluctuations in IT
workloads. In this regard, a client company can
leverage a SaaS vendor's capacity to adapt to change.
Focus on core competencies: SaaS adoption will
also facilitate firms' refocusing on their core
competences. This refocusing is possible by
84
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
all customers at once, which implies a high value at
risk [13].
Integration risk: risk of problems related to the SaaS
application's interoperability and integration with
homegrown applications located on the client side.
Potential losses due to performance risks can be
significant because the day-to-day operations will not
be optimally supported [12].
Other Cases
Working experiences (Year)
1~3
3~5
5~10
>10
9(14%)
19(29%)
29 (45%)
8(12%)
We measured the validity by factor analysis test and
reliability by Cronbach's alpha test. According to the
results of factor analysis to test kmo = 0.7 and α = 0.8
for the alpha test has demonstrated high reliability
and validity. The questionnaire is designed in likert
spectrum. We evaluate the normal distribution of the
data by the Kolmogorov – Smirnov test. So according
to the test results, the values of significance level all
components are less than 0/05, so the non-normal
data distribution. Therefore, in this study the
binomial test was done (Sign-level=0.05, cut
point=3) by SPSS software. So the research
hypothesis is as follow:
5. Analysis Methodology
To analyze the factors identified in the previous
section, a questionnaire among 192, employees and
managers of government agencies and private
companies distributed that were the 65 questionnaires
were answered with a success. Frequency distribution
of respondents' profile in table 1 is shown.
Table 1: Frequency distribution of respondents' profile
Characteristic
Type of Organization
Governmental organizations
Private organizations
Roles
Organization manager
IT manager
Computer and IT engineer
R&D
2(3%)
Sample composition
Hypotheses A: Understanding and knowledge of
cloud computing and SaaS.
HA-1: Respondents' awareness of cloud computing is
desirable.
HA-2: Respondents' awareness of SaaS is desirable.
Due to the novelty of cloud computing, in this case
we
determine
cut
point
on
2
24(37%)
41(63%)
4(6%)
15(23%)
36(56%)
8(12%)
Table 2: The results of testing HA hypothesis (sig. level 0.05)
Awareness
Cloud
Awareness
Group
1
2
Total
N
Binomial test
Observed
Category
Prop.
15
<= 2
.23
50
65
>2
.77
.50
Exact Sig.
(2-tailed)
Reject/confirm
the hypothesis
.000
Confirm
1.00
SaaS
1
24
<= 2
.37
Awareness
2
41
65
>2
.63
Total
Test
Prop
.50
.046
Confirm
1.00
According to table 2 the values of significance level
HA-1 and HA-2 is less than 0.05 and the frequency of
observations category (>2) is more, so they are
confirmed. Thus, according to all frequency of
observations, respondents have confirmed hypothesis
A.
Hypotheses B: Respondents believe that the SaaS
has many benefits.
HB-1: The SaaS is accessible anywhere at any time.
HB-2: IT infrastructure is not needed for SaaS
utilization.
HB-3: SaaS utilization reduce software support and
data management tasks dramatically.
HB-4: SaaS utilization is lead to reduce costs.
85
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Table 3: The results of testing HB hypothesis (sig. level 0.05)
Benefits
Group
N
Category
Binomial test
Observed
Test
Prop.
Prop.
Access
1
24
<= 3
.37
Anywhere
2
41
>3
.63
Total
65
1
40
<= 3
.62
infrastruct
2
25
>3
.38
ure
Total
65
Software
1
46
<= 3
.71
Maintenan
2
19
>3
.29
ce
Total
65
Lower cost
1
20
<= 3
.31
2
45
>3
.69
Total
65
Zero
IT
.50
Exact Sig. (2tailed)
Reject/confirm
the hypothesis
.046
Confirm
1.00
.50
.082
Reject
1.00
.50
.001
.50
.003
Reject
1.00
Confirm
1.00
According to table 3 the values of significance level
HB-1 and HB-4 are less than 0.05 and the frequency of
observations category (>3) is more, so it is
confirmed. The values of significance level HB-3 is
less than 0.05 but the frequency of observations
category (<= 3) is more, so it is rejected. The values
of significance level H B-2 is more than 0.05, so it is
rejected. Thus, According to all frequency of
observations, respondents have rejected hypothesis B.
Hypotheses C: Respondents believe that the SaaS
has many Weaknesses.
Hc-1: SaaS still has not matured in Iran.
Hc-2: SaaS pricing is Ambiguous and complicated.
Hc-3: SaaS utilization is leads to Dependency on the
Provider.
Hc-4: cloud is Completely Dependent on the internet.
Table4: The results of testing HC hypothesis (sig. level 0.05)
Weaknesses
Group
N
Binomial test
Observed
Category
Prop.
Immature
1
22
<= 3
.34
SaaS
2
43
>3
.66
Total
65
complex
1
39
<= 3
.60
pricing
2
26
>3
.40
Total
65
Dependency
1
42
<= 3
.65
on the provide
2
23
>3
.35
Total
65
Dependency
1
24
<= 3
.37
on the Internet
2
41
>3
.63
Total
65
Test
Prop.
.50
Exact Sig.
(2-tailed)
Reject/confirm
the hypothesis
.013
Confirm
1.00
.50
.136
Reject
1.00
.50
.025
Reject
1.00
.50
.046
Confirm
1.00
86
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
According to table 4 the values of significance level
HC-1 and HC-4 are less than 0.05 and the frequency of
observations per category (>3) is more, so they are
confirmed. The values of significance level HC-3 is
less than 0.05 but the frequency of observations
category (<= 3) is more, so it is rejected. The values
of significance level HC-2 more than 0.05, so it is
rejected. Thus, According to all frequency of
observations, respondents have rejected hypothesis C.
Hypotheses D: Respondents believe that SaaS
adoption brings many opportunities.
HD-1: SaaS adoption is leads to saving the cost.
HD-2: SaaS adoption provides a great degree of
flexibility.
HD-3: SaaS adoption will also facilitate firms'
refocusing on their core competences.
HD-4: SaaS adoption Brings Access to specialized
resources.
Table 5: The results of testing HD hypothesis (sig. level 0.05)
Opportunities
Cost saving
Group
N
Category
Binomial test
Observed
Test
Prop.
Prop.
1
23
<= 3
.35
2
42
>3
.65
Total
65
Strategic
1
24
<= 3
.52
flexibility
2
41
>3
.48
Total
65
1
57
<= 3
.88
core
2
8
>3
.12
competencies
Total
65
Access
1
46
<= 3
.71
specialized
2
19
>3
.29
resources
Total
65
Focus
on
to
.50
Exact Sig. Reject/confirm the
hypothesis
(2-tailed)
.025
Confirm
1.00
.50
.804
Reject
1.00
.50
.000
Reject
1.00
.50
.001
Reject
1.00
According to table 5 the values of significance level
HD-1, is less than 0.05 and the frequency of
observations category (>3) is more, so it is
confirmed. The values of significance level HD-3 and
HD-4 are less than 0.05 but the frequency of
observations per category (<= 3 ) is more, so they are
rejected. The values of significance level HD-2 more
than 0.05, so it is rejected. Thus, According to all
frequency of observations, respondents have rejected
hypothesis D.
Hypotheses E: Respondents believe that SaaS
adoption has many risks.
HE-1: SaaS adoption has many lack of control risks.
HE-2: SaaS adoption has many legal risks.
HE-3: SaaS adoption has many security risks.
HE-4: SaaS adoption has many Performance risks.
HE-5: SaaS adoption has many economic risks.
HE-6: SaaS adoption has Risks related to Internet
resilience and bandwidth.
HE-7: SaaS adoption has risk related to problem of the
application's integration.
87
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Table6: The results of testing HE hypothesis (sig. level 0.05)
Risks
Lack
Group
of
Binomial test
Observed
Test
Category
Prop.
Prop.
N
1
24
<= 3
.37
2
41
>3
.63
Total
65
Legal
1
36
<= 3
.55
issues
2
29
>3
.45
Total
65
1
17
<= 3
.26
2
48
>3
.74
Total
65
Performanc
1
57
<= 3
.88
e
2
8
>3
.12
Total
65
1
54
<= 3
.83
2
11
>3
.17
Total
65
1
24
<= 3
.37
2
41
>3
.63
Total
65
1
23
<= 3
.35
2
42
>3
.65
Total
65
control
Security
Economic
bandwidth
Integration
.50
Exact Sig. Reject/confirm
the hypothesis
(2-tailed)
.046
Confirm
1.00
.50
.457
Reject
1.00
.50
.000
Confirm
1.00
.50
.000
Reject
1.00
.50
.000
Reject
1.00
.50
.046
Confirm
1.00
.50
.025
Confirm
1.00
the analysis of organizations understand of the benefits,
weaknesses, opportunities and risks of SaaS. According to
According to table 6 the values of significance level
HE-1, HE-3, HE-6 and HE-7 are less than 0.05 and the
frequency of observations per category (>3) are
more, so all is confirmed. So values of significance
level HE-4 and HE-5 are less than 0.05 but the
frequency of observations per category (<= 3) are
more, so they is reject. The values of significance
level HE-2 more than 0.05, so it is rejected. Thus,
According to all frequency of observations,
respondents have rejected hypothesis E.
the results, Respondents believe that the more main
benefits of the SaaS are reducing the cost and
permanent availability. Also leading to savings in the
overall cost of organization. Immature SaaS and
Dependency on the Internet confirmed as a more main
Weaknesses of SaaS. They are most concerned about
Security and lack of control risks and risk related to
problem of the application's integration and Internet
resilience and bandwidth. Iranian organization
perceived more SaaS risks than SaaS benefits. Thus,
they are not tended to adopt SaaS. Emergence of
successful SaaS business models can help to adopting
the SaaS.
6. Discussion and conclusions
Software as a Service (SaaS) is a relatively new
organizational application sourcing alternative,
offering organizations the option to access
applications via the Internet. We in study focuses on
88
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
[10] A. Benlian and T. Hess and P. Buxmann, “Drivers of
SaaS-adoption: an empirical study of different application
types”, Business & Information Systems Engineering 1 (5),
2009, 357–369
[11] R Narwal and S Sangwan, “Benefits, Dimensions and
Issues of Software as a Service (SAAS)”, International
Journal of New Innovations in Engineering and
Technology (IJNIET), 2013,
[12] A. Benlian and T. Hess, “Opportunities and risks of
software-as-a-service: Findings from a survey of IT
executives”, Decision Support Systems 52, 2011, 232–246
[13] R.J. Kauffman and R. Sougstad, “Risk management of
contract portfolios in IT services: the profit-at-risk
approach”, Journal of Management Information Systems 25
(1), 2008, 17–48.
[14] H. Gewald and J. Dibbern, “Risks and benefits of
business process outsourcing: a study of transaction
services in the German banking industry”, Information and
Management 46 (4), 2009, 249–257
[15] S. Garg and S. Versteeg and R. Buyya, “A framework
for ranking of cloud computing services”, 2012, Future
Generation Computer Systems
[16] W. Wu and L.W. Lan and Y Lee, “Exploring decisive
factors affecting an organization’s SaaS adoption: A case
study”, International Journal of Information Management
31 (2011) 556–563
[17] We. Wu,” Developing an explorative model for SaaS
adoption”, Expert Systems with Applications 38 (2011)
15057–15064
[18] M. Madisha, “Factors Influencing SaaS Adoption by
Small South African Organisations1”, Conference on
World Wide Web Applications, Durban, South Africa,
2011
Reference
[1] G. Feuerlicht, “Next generation SOA: Can SOA survive
cloud computing?”, In: V. Snasel P.S. Szczepaniak, & J.
Kacprzyk, (Eds.), Advances in Intelligent Web Mastering 2, AISC 67, 2010, pp. 19–29.
[2] D. Catteddu and G. Hogben, “Cloud computing:
Benefits, risks and recommendations for information
security”, European Network and Information Security
Agency (ENISA), 2009, 1–125.
[3] N. Sultan,” Cloud computing for education: A new
dawn?” International Journal of Information Management,
2010, 30(2), 109–116.
[4] S. Subashini and V. Kavitha,” A survey on security
issues in service delivery models of cloud computing”,
Journal of Network and Computer Applications, 2010, 111.
[5] A. Susarla and A. Barua, “Multitask agency, modular
architecture, and task disaggregation in SaaS”, Journal of
Management Information Systems, 26, 2010, 87–118.
[6] S. Marstson, and Z. Li, “Cloud computing – The
business perspective”, Decision
Support Systems, 51, 2011, 176–189
[7] V. Choudhary, “Comparison of software quality under
perpetual licensing and software as a service”, Journal of
Management Information Systems, 24, 2007, 141–165.
[8] L. Wu, S and Garg and R. Buyya, “SLA-based
admission control for a Software-asa-Service provider in
Cloud computing environments”, Journal of Computer and
System Sciences, 78(5), 2012, 1280–1299
[9] Gartner, “Forecast analysis: Software as a service”,
worldwide,
2010b,http://www.scribd.com/doc/49001306/ForecastAnalysis-Software
as-a-Service-Worldwide-2009-2014.
Accessed 10.11.11 [Online].
89
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Evaluation of the CDN architecture for Live Streaming through
the improvement of QOS
Case study: Islamic Republic of Iran Broadcasting
Shiva Yousefi1, Mohamad Asgari2
1
Taali university,
Ghom, Iran
[email protected]
2
Islamic Republic of Iran Broadcasting university,
Tehran, Iran
[email protected]
- High traffic, packet loss and narrow channels of
broadcast decrease the quality of the content significantly.
In light of the discussed issues and the endless growth
of Internet usage in the past few years, providing services
such as Live Television programs, On-demand clips and
music, and social networks such as Facebook and
YouTube simultaneously to thousands of users has
become a challenge. CDN network providers use
companies as a tool in terms of sharing the content at the
discretion of the user. Reduce in the video quality and lack
of access associated mainly resulting from the long
loading (download) can be tiring and boring for the users.
Companies make great profit from e-commerce on the
Internet, and would like to create a suitable experience for
visitors to their website. Therefore, in recent years, the
goal to improve the quality of technology for sending
content through the Internet has grown dramatically. The
content networks try to increase the quality of services by
using different mechanisms. But unfortunately no
company in Iran has invested in this field yet. The only
Web site that is located on the CDN is live.irib.ir, which is
owned by the Islamic Republic of Iran's Broadcasting and
it uses a few CDNs in foreign countries to transfer the
content and its central server inside the same country
offers services to users inside the country. A CDN is a
collaborative collection of network elements spanning the
Internet, where content is replicated over several mirrored
Web servers in order to perform transparent and effective
delivery of content to the end users [1]. By maximizing
bandwidth, increasing accessibility and repeating content,
CDN's reduce the response time to the user’s request, and
therefore lead to greater network productivity. It is
predicted that by employing the technology in our country,
we will witness better network performance in content
delivery.
This paper proposes the architecture of CDN in Iran with
Content Re-distribution Function to overcome the
limitations of watching live streaming in the Internet, and
Abstract
Over the last decades, users have witnessed the growth and
maturity of the Internet. As a consequence, there has been an
enormous growth in network traffic, driven by rapid acceptance
of broadband access, along with increases in system complexity
and content richness. CDN, provides services that increases the
efficiency of the network bandwidth, and increases the access
and maintains the accuracy of the information during
transmission. One of the most important requests in the context
of the Internet is watching video live streaming with techniques
that can be transferred. The aim of this study is to enhance the
QOS and increase the efficiency of the transfer of the content in
the context of the Internet. The results of this research in the form
of diagrams in order to compare the advantages and
disadvantages of using a CDN will be shown at the end of the
review proposals. Finally we concluded that the use of this
technology in the context of the Internet in Iran can increase the
QOS and customer satisfaction levels.
Keywords: CDN, Live Streaming, QOS, Delay, Jitter, Packet
loss, surrogate.
1. Introduction
Limited access to the network and computational resources
can become the arena for competing service centers with
one another. High availability and optimal answers
provided the key to success in this arena. At the same time
the influx of high volumes of users to a Web site and
creation of a high traffic, became one of the challenges of
this the realm. Some of the problems resulting from this
situation are as follows:
- Since content must go through several long distances and
backbone passes for the relocation the transmission speed
is going to greatly decreased.
- Traffic along with issues related to backbone peering,
causes issues with the proper delivering of content.
- The parent site's bandwidth limits the usage of the users.
90
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
it explains about QOS factors and clients satisfaction
based on Content Delivering Network (DCDN).
In the rest of the sections, first we discuss the architecture
of CDN, live streaming and Islamic Republic of Iran
Broadcasting’s website.
Then we introduce the proposed CDN architecture in Iran,
content
re-distribution
algorithm
and
RDCDN
transmission algorithm that make a more stable
transmission performance to clients. Finally we present the
simulation results which prove our proposed model, and
the conclusion.
2. Content Delivery Network Architecture
Figure1: CDN node distribution over the internet
Content Delivery Network (CDN) is one of the new
popular technologies which are used for distributing
content on Internet. With this technology, corporations and
institutions can increase the speed of uploading and
browsing their sites. As you know, faster uploading site to
optimize the site for search engines (SEO) is currently
very important in terms of e-commerce and also has
positive results.
Internet Corporations for controlling traffic of the sites and
Internet services use multiple servers around the world for
storing and distributing their information. This makes it
possible for all users around the world to have the same
condition while using this service and using the closest
server of corporation with highest speed. This will bring
many advantages that will be described in the following
sections but what about the smaller companies and the
Internet users who have shared their personal sites on the
Internet?
Usually these sites keep their data on a single server (their
host) and service the users and the visitors around the
world and carry out their work with their server limitation
such as low bandwidth, low transmission speed and etc. [1]
CDN can be efficient and help small business owners and
private sites to act like corporations. This service keep a
copy of downloadable information of your site (like CSS
code files, java scripts code files, multimedia files, etc) on
different nodes or servers around the world and service
many sites at the same time. At this time when client want
to browse your site, information will deliver from the
closest available server to him and because these
information were already cached, exchange rate has
changed dramatically.
Figure1 can help to better understand the concept of CDN
in the world of internet [2].
3. Functions and Structure of CDN
As mentioned before, CDN is a technology with its
Internet corporations control traffic of the sites and
Internet services. With this technology users can connect
to the nearest server and achieve their desired content.
For further explanation about CDN’s construction, some
concept must be defined:
3.1 Surrogate server
CDN consist of several edge servers (surrogate) which
keep a copy of provider’s contents so they called cache
servers too. Client’s requested content will deliver to
nearest surrogate server. Every cluster consists of set of
surrogates which are located in same geographical region.
The number of surrogates is too important because they
should be in appropriate number and also they should be in
appropriate location [2].
3.2 Content server
Content server: content providers upload client’s requested
content in content servers.
CDN is a set of network elements which combine together
to deliver a content to clients in a more efficient way.
A CDN provides better performance through caching or
replicating content over some mirrored Web servers (i.e.
surrogate servers) strategically placed at various locations
in order to deal with the sudden spike in Web content
requests, which is often termed as flash crowd [3] or
SlashDot effect [4]. The users are redirected to the nearest
surrogate server to them. This approach helps to reduce the
network impact on the response time of user requests. In
the context of CDNs, content refers to any digital data
resources and it consists of two main parts: the encoded
media and metadata [5]. The encoded media includes static,
91
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
dynamic and continuous media data (e.g. audio, video,
documents, images and Web pages). Metadata is the
content description that allows identification, discovery,
and management of multimedia data, and also facilitates
the interpretation of multimedia data.
3.3 Components of CDN
The three key components of a CDN architecture are
content provider, CDN provider and end-users. A content
provider or customer is one who delegates the URI name
space of the Web objects to be distributed. The origin
server of the content provider holds those objects. A CDN
provider is a proprietary organization or company that
provides infrastructure facilities to content providers in
order to deliver content in a timely and reliable manner.
End-users or clients are the entities who access content
from the content provider’s website.
CDN providers use caching and/or replica servers located
in different geographical locations to replicate content.
CDN cache servers are also called edge servers or
surrogates. The surrogates of a CDN are called Web
cluster as a whole. CDNs distribute content to the
surrogates in such a way that all cache servers share the
same content and URL. Client requests are redirected to he
nearby surrogate, and a selected surrogate server delivers
requested content to the end-users. Thus, transparency for
users is achieved. Additionally, surrogates send accounting
information for the delivered content to the accounting
system of the CDN provider [1].
Figure 2 shows a typical content delivery environment
where the replicated Web server clusters are located at the
edge of the network to which the end-users are connected
[1]. A content provider (i.e. customer) can sign up with a
CDN provider for service and have its content placed on
the content servers. The content is replicated either ondemand when users request for it, or it can be replicated
beforehand, by pushing the content to the surrogate servers.
A user is served with the content from the nearby
replicated Web server. Thus, the user ends up
unknowingly communicating with a replicated CDN
server close to it and retrieves files from that server.
CDN providers ensure the fast delivery of any digital
content. They host third-party content including static
content (e.g. static HTML pages, images, documents,
software patches), streaming media (e.g. audio, real time
video), User Generated Videos (UGV), and varying
content services (e.g. directory service, e-commerce
service, file transfer service). The sources of content
include large enterprises, Web service providers, media
companies and news broadcasters. The end-users can
interact with the CDN by specifying the content/service
request through cell phone, smart phone/PDA, laptop and
desktop [1].
Figure 2: Abstract architecture of a Content Delivery Network (CDN)
Figure 3 depicts the different content/services served by a
CDN provider to end-users [1].
Figure 3. Content/services provided by a CDN
4. Live Streaming
Streaming media delivery is challenging for CDNs.
Streaming media can be live or on-demand. Live media
delivery is used for live events such as sports, concerts,
channel, and/or news broadcast. In this case, content is
delivered ‘instantly’ from the encoder to the media server,
and then onto the media client. In case of on-demand
delivery, the content is encoded and then is stored as
streaming media files in the media servers. The content is
available upon requests from the media clients. Ondemand media content can include audio and/or video ondemand, movie files and music clips [1].
Streaming means sending data and this data usually is
sound or video. It let users play the stream before
receiving and processing the full stream and the live
streaming protocols are used for control the transmission.
4.1 Components of Live Streaming
Here are the components of live streaming [6]:
- Video sources that produce the video.
- A Live encoder is a hardware/software capable of
compressing the video source, in real-time, and send the
92
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Number of bits (packet length) and “R” is the Rate,
(bandwidth) in bits per second.
compressed video to a media server in a specific format,
using a specific protocol.
- The media server is a software that installs on a dedicated
server and it is optimized/specialized in "serving" media
content to the end users, through a WebTV (for example).
The media server takes the compressed video supplied by
the live encoder and broadcast it to the users.
- The players that connect to media server and play the
content of media to clients.
Dt = N/R
5.2 Jitter
Jitter refers to the changes in delay, which is calculated
through the variations in the length of delays in the
network. A network with a constant delay rate, has no
Jitter (calculated with Formula 4).
5. Quality of Service in CDN
Delay jitter = Maximum time – Minimum delay time
One major goal of CDN and in fact one of the reasons for
its creation is increasing quality of service and customer
satisfaction levels. The most important factors of
increasing Quality of Service (QOS) are reducing Delay,
Jitter and Packet loss. For a better understanding we will
explain each of these concepts [7].
(4)
5.3 Packet loss
Packet loss means the loss of packet through transmission.
One of the reasons that packet loss occurs is that the
uplinks and switches of the network send more packets
than the buffer packets; and another reason is the loss of
connection. The easiest way to calculate packet loss is
using the ping command [9].
5.1 Delay
The time that a packet need to be transmit from the source
to the destination in a network. It is calculated according to
the following formula:
6. Islamic Republic of Iran’s Web Site
Islamic Republic of Iran’s Web Site, with the address
live.irib.ir, is the only web site in the country in the
context of CDN that has shared its contents around the
world. A content distribution network makes it possible
for users in remote out of coverage areas to TV channels
and radio broadcasting stations via their websites accessed
by personal computers or mobile devices. This service
must have access to the image files of the organization
comply with security tips with the contacts to people
around as well. The Islamic Republic of Iran Broadcasting
has 6 main servers. One of the servers is in USA, two of
them in Europe and also three servers in Asia and it as
thirty surrogated servers which are scattered all over the
continents.
End-to-end delay=N[transmission delay + propagation
delay + processing delay]
(1)
In this formula “N” is the number of routers + 1
5.1.1 Processing delay
The time that the router needs to process the header of the
packet to determine the next path. Processing delay in the
network is usually measured in microseconds or even less
[9].
5.1.2 Propagation delay
It shows the amount of time it takes for the head signal to
reach the receiver from the sender
It will be calculated with the formula (2), where “d” is the
length of the physical link and “s” is velocity in the
medium
Propagation delay = d/s
(3)
7. Proposed CDN
In the short time that has passed since the birth of ecommerce in today's world, it has seen unprecedented
growth in developing countries, and is predicted to expand
enormously on a global scale in the near future. One of the
businesses that large foreign countries have heavily
invested into these days is providing CDN's, which has
been described in some detail in previous sections.
Unfortunately in our country so far, the company has not
attempted to do this. Perhaps because of its multiple
servers setup in money capital from all parts of the country,
or it may be because of its high weakness of our
telecommunication infrastructure or maybe because of
(2)
5.1.3 Transmission delay
The amount of time needed for all data packets in the wire
to be sent. In other words, this delay is the given rate in the
link. Transmission rate is directly related to the length of
the packet, and is independent of the distance between the
two nodes. This delay is calculated with Formula (3), in
which “Dt” is Delay Time in seconds and “N” is the
93
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
uncertainty of getting this technology in the country. But
given the very rapid progress of technology in our country,
the existence of such technology will soon be needed and
even essential. The establishment of virtual universities in
the country is another advantage of advancing
technologies, and as was mentioned in the previous
sections, the spread of CDN in the country can be a great
help to universities and scholars in increasing the quality
of conduct for online classrooms. Today, using the internet
to watch movies or listen to radios or other content sources
has become ubiquitous, and as mentioned earlier, IRIB is
the one and only content provider that uses CDN
technology for distributing their media. This institution,
due to the lack of CDN services in our country, resorts to
using CDN from other countries. This is only ideal if a
user requests content on the IRIB's website from another
country, in which case they will be connected to the
nearest server. If someone from Southern Iran requires
access to this content, they must connect to the server in
Tehran and receive their desired content despite all the
difficulties and barriers, since it is all stored in a single
server located in Tehran. However, if there was a CDN in
the country and the content of this website were stored on
the CDN's servers, this user could have connected to the
nearest server containing the data, and had quicker and
higher quality access to their desired content. In the next
section, through the experiment that will be conducted, we
will discuss whether the given suggestion will result in
increased quality of service for customers or not,
considering the aforementioned factors; and this
suggestion will be tested on the server that IRIB's website
is stored on.
Figure 4. The Proposed Architecture of Iran CDN
This architecture is modeled from RDCDN architecture
that was introduced in 2009 [8].
In this architecture, main surrogate can operate as a
backup storage system controlled by local server. So, even
though sub-surrogates delete their stored contents in the
storage
spaces, the clients can continuously receive contents
services from main surrogate, and sub-surrogates can
receive the deleted contents from main surrogate to
recover them without re-transmission of content provider.
8.1 Content distribution Algorithm
Content redistribution algorithm transfers contents to each
sub-surrogate according to the features based on request
rates of contents during the set-up time. Proposed content
distribution algorithm manages all contents by grouping,
and each sub-surrogate receives one content per groups at
least. At that time, sub-surrogate which has the content
that is the highest request rate in its group will receive the
lowest content of next group [8].
8. Simulating proposed CDN in Iran
Now we simulate a CDN in Iran. The main elements of
CDN architecture are [8]:
Content providers (in our simulation Islamic Republic of
Iran’s broadcasting), main servers and surrogates including
main surrogates and sub-surrogates, and they are arranged
in hierarchical order as in Figure 4.
94
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Figure 6. CDN Server’s physical location arrangement
For testing and calculation the conclusions, we consider
one of the users that is located in the city of Shiraz (Fars
province) in this article. Due to being closer to the
Isfahan's server, it appears the connection to the Isfahan's
server the user delay, takes less jitter and Packet loss than
when the server is connected directly to Tehran. To ensure
this issue, we connect to servers in both cities and we
compare the listed parameters. To do this, firstly we send
the packages with the size of 32 bytes and then the
package with the size of 1024 bytes to Tehran and Isfahan
servers and we extract the results. After the three main
calculated factors of QOS, in Figure (7) we can see the
differences of the delay when the user is closer to the
nearest server (Isfahan) than to the main server (Tehran).
Figure (8) can show the differences in connection of jitter
when the user is closer to the nearest server (Isfahan) and
the main server (Tehran).
Figure 5. The Flow Diagram of Content distribution Algorithm
Figure (5) shows the flow diagram for content distribution
algorithm. By this algorithm, sub-surrogates maintain their
prior storing states if priority of contents is changed but
group of content is not changed. Therefore, entire network
load is decreased [8].
8.2 Simulating and results
To simulate the implementation of CDN in Iran and also to
prove the efficiency of the offer, we consider the
surrogated servers in three big cities of Iran, which are
located in three different geographic areas. For example,
surrogated servers are located in the cities of Tabriz,
(North West), Mashhad (North-East), Isfahan (South). The
main server (Media Server) of the Broadcast of Iran is
located in the city of Tehran. According to the
consideration of first factor in the CDN in transferring
traffic, which is choosing closest server to the requested
user, therefore, is expected for users to connect to the
nearest server at the time of the request. In order to
demonstrate the improvement in user experience in
implementing this suggestion, a few users in different
cities send packets to new servers and Tehran's server,
which allows us to draw a conclusion and compare the
new method with the person old methods using the QOS
parameters. The CDNs server's physical views are shown
in the figure 6.
Figure 7. Delay of sending packets from Shiraz to Tehran and Isfahan’s
servers
95
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
companies that are the founder of this technology in Iran,
in addition to the benefits it has, they can make some kind
of type and foster entrepreneurship in the country, and by
that a very large amount of jobs can be created and also
the companies can use from college graduates and
professionals in the field. The above research can be
extended and tested by more users in the future to ensure
the function for the accuracy achieved.
jitter
25.000
20.000
15.000
10.000
5.000
0.000
32 byte
1024 byte
32 byte
1024 byte
Shiraz
Shiraz
Shiraz
Shiraz
Isfahan
Isfahan
Tehran
Tehran
References
[1] M. Pathan, and R. Buyya, and A. Vakali, Content Delivery
Network, Springer-Verlag Berlin Heidelberg, 2008.
[2] J. Mulerikkal, and I. Khalil, "An Architecture for Distributed
Content Delivery Network" , IEEE Network, 2007, pp.359364.
[3] M. Arlitt, and T. Jin, "A Workload Characterization Study of
1998 World Cup Web Site", IEEE Network, 2000, pp. 30-37.
[4] S. Adler, "The SlashDot Effect: An Analysis of Three
Internet Publications", Linux Gazette Issue, Vol. 38, 1999.
[5] T. Plagemann, and V. Goebel, and A. Mauthe, L. Mathy, and
T. Turletti, and G. Urvoy-Keller, "From Content Distribution
to Content Networks – Issues and Challenges" Computer
Communications, Vol. 29, No. 5, 2006, pp. 551-562.
[6] WebTV Solutions, http://www.webtvsolutions.com/, 2013.
[7] N. Bartolini, and E. Casalicchio, and S. Tucci, "A Walk
Through Content Delivery Networks", In Proceedings of
MASCOTS 2003, LNCS Vol. 2965/2004, 2004, pp. 1-25.
[8] M. Sung, and Ch. Han, "A study on Architecture of
CDN(Content Delivering Network) with Content Redistribution Function" , IEEE Network, 2009, pp.727-777.
[9]
University
of
Minnesota
Duluth,
http://
http://www.d.umn.edu/, 2013.
Figure 8. Jitter of sending packets form Shiraz to Shiraz and Isfahan’s
servers
Figure 9 shows the difference of connection of Packet loss
when the user is closer to the nearest server (Isfahan) and
the main server (Tehran) can be seen.
Paket loss
40.000
30.000
20.000
10.000
0.000
32 byte
1024 byte
32 byte
1024 byte
Shiraz
Shiraz
Shiraz
Shiraz
Isfahan
Isfahan
Tehran
Tehran
Shiva Yousefi received the associate degree from DPI college in
Tehran, B.S. degree from Iran Info-Tech department university in
Tehran and M.S. degree from Taali university in Qom, all in
computer sciences. She is currently working as a Network
Administrator in Islamic Republic of Iran Broadcasting. She is
involved in research and development project about CDN in
Islamic Republic of Iran broadcasting.
Figure 9. Packet loss of sending packets from Shiraz to Tehran and
Isfahan’s servers
This simulation has been tested on a few other cities and
other users and the extracted results were almost the same.
After reviewing these figures, it can be concluded that
when the users after their request, they connect to the
nearest server for receiving their content, the amount of
delay of their Jitter and Packet loss would be significantly
reduced.
Mohamad Asgari received the P.H.D. degree from Iran University
of Science and Technology in Tehran. He is currently working as
director of the department of development and technology in
Islamic Republic of Iran broadcasting and also He is professor in
IRIB broadcasting university. He is involved in research and
development project about CDN in Islamic Republic of Iran
broadcasting.
9. Conclusion and Future work
According to the simulation, after doing the calculations, it
can be concluded that the existence of CDN in Iran, when
users request their content according to the existing
algorithm that were introduced earlier would connect to
the nearest server and consequently the amount of delay,
Jitter and Packet loss will be reduced significantly. Items
listed are the three basic calculations of the items of the
quality of the services. So it can be said that this would
enhance the quality of the services. The establishment of
96
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
A Robust and EnergyEnergy-Efficient DVFS Control Algorithm
for GALSGALS-ANoC MPSoC in Advanced Technology
under Process Variability Constraints
1,∗Hatem Zakaria2, †Laurent Fesquet3 ‡
∗ 1
Sylvain
Nicolas
Marchand
SylvainDurand
Durand
, Hatem Zakaria , Laurent Fesquetand
, Nicolas
Marchand
1
Control Department of GIPSA-lab, CNRS, Univ. of Grenoble
Control Department of GIPSA-lab, CNRS, University of Grenoble
Grenoble,
France
Grenoble, France
[email protected], [email protected]
†
Electrical Engineering Department, Benha Faculty of Engineering, Benha University
Egypt
2 Electrical Engineering Department,Benha,
Benha Faculty of Engineering, Benha University
‡
TIMA Laboratory,
Benha, Grenoble
Egypt University
Grenoble, France
[email protected]
E-mail: [email protected]
3
TIMA Laboratory, Grenoble University
Grenoble, France
Abstract
tainty about how a fabricated system will perform [1]. [email protected]
though a circuit or chip is designed to run at a nominal clock
Many Processor Systems-on-Chip (MPSoC) have become tremenfrequency, the fabricated implementation may vary far from
dously complex systems. They are more sensitive to variability
unevenness, and changeability associated with a given
with technology scaling, which complicates the system design and
this expected performance. Moreover, some cores may beAbstract
feature or speci-fication. It has became one of the leading
impact the overall performance. Energy consumption is also of
have differently inside the same chip [2]. To solve these
Many
Processor
Systems-on-Chip
(MPSoC)
haveandbecome
causes for chip failures and delayed schedules at a subgreat interest
for mobile
platforms powered
by battery
power
problems, MPSoCs in advanced technology require a dytremendously
complex systems.
They onareDynamic
more sensitive
to
management techniques,
mainly based
Voltage and
micrometric scale and complicates system design by
namic power management in order to highly reduce the envariability
technology
scaling,
which become
complicates
the system
Frequencywith
Scaling
(DVFS)
algorithms,
mandatory.
A
introducing uncertainty about how a fabricated system will
ergy consumption. Architectural issues are also needed for
Globally
Locally
Synchronous
(GALS)
design aldesign
andAsynchronous
impact the overall
performance.
Energy
consumption
perform [1]. Although a circuit or chip is designed to run
problems
clocks,
each by
onebattery
being
isleviate
also ofsuch
great
interest by
forhaving
mobilemultiple
platforms
powered
helping the yield enhancement of such circuits with strong
at a nominal clock frequency, the fabricated
distributed
on
a
small
area
of
the
chip
(called
island),
whereas
and power management techniques, mainly based on Dynamic
technological uncertainties.
implementation may vary far from this expected
an
Asynchronous
Network-on-Chip
(ANoC)
allow
to
communiVoltage and Frequency Scaling (DVFS) algorithms, become
Whereas
leakageMoreover,
power is a some
significant
contributor
the
cate betweenAtheGlobally
different islands.
A robustLocally
technique
is proposed
performance.
cores
may to
behave
mandatory.
Asynchronous
Synchronous
to deal with a GALS-ANoC architecture under process variabiltotal
power,
the
average
power
consumption
and
the
energy
differently inside the same chip [2]. To solve these
(GALS) design alleviate such problems by having multiple
ity constraints using advanced automatic control methods. The
dissipation
dominated
by the dynamic
power inrequire
current a
clocks,
each
one
being
distributed
on
a
small
area
of
the
chip
problems, are
MPSoCs
in advanced
technology
approach relaxes the fabrication constraints and help to the yield
embedded
CMOS
integrated
circuits.
The
energy
reduction
(called
island),
whereas
an
Asynchronous
Network-on-Chip
dynamic power management in order to highly reduce the
enhancement. Moreover, energy savings are even better for the
is
a quadratic
functionArchitectural
of the voltageissues
and a are
linear
function
(ANoC)
allow toperformance
communicate
between
the different
islands.
A
energy
consumption.
also
needed
same perceived
with
the obtained
variability
robustrobust
technique
is proposed
to based
deal on
with
a GALS-ANoC
of
the
clock
frequency
[3].
As
a
result,
Dynamic
Voltage
ness. The
case study
is an island
a MIPS
R2000 profor helping the yield enhancement of such circuits
with
architecture
under process
variability constraints
advanced
cessor implemented
in STMicroelectronics
45nm using
technology
and
Scaling
(DVS) can be
used to efficiently manage the enstrong technological
uncertainties.
automatic
methods. simulations.
The approach relaxes the fabrication
validated control
with fine-grained
ergy consumption of a device [4]. Supply voltage can be
Keywords:
control,
power MPSoC,
process
variconstraints
andPredictive
help to the
yield low
enhancement.
Moreover,
energy
reduced
slack isisavailable
in the contributor
critical path,tobut
Whereaswhenever
leakage power
a significant
the
ability are
robustness,
DVFS,
ANoC.
savings
even better
forGALS,
the same
perceived performance with
one
has
to
take
care
that
scaling
the
voltage
of a microprototal power, the average power consumption
and the
the obtained variability robustness. The case study is an island
cessor
its speed
as well. by
Therefore,
adapting
thein
energy changes
dissipation
are dominated
the dynamic
power
based on a MIPS R2000 processor implemented in
supply
voltage
is
very
interesting
but
implies
the
use
of
Dy1.
Introduction
STMicroelectronics 45nm technology and validated with finecurrent embedded CMOS integrated circuits. The energy
namic
Frequency
Scalingfunction
(DFS) toofkeep
correct the
grained simulations.
reduction
is a quadratic
the voltage
andsystem
a linear
behavior.
The
addition
of
DFS
to
DVS
is
called
The upcoming
generations
of integrated
Many
Keywords:
Predictive
control,
low powersystems,
MPSoC,asprocess
function of the clock frequency [3]. As a result,Dynamic
Dynamic
Voltage
Frequency
and resultsmanage
in siProcessorrobustness,
Systems-on-Chip
(MPSoC),
variability
DVFS, GALS,
ANoC.require drastic techVoltage and
Scaling
(DVS) Scaling
can be (DVFS)
used to efficiently
∗
nological evolution since they have reached limits in terms
of power consumption, computational efficiency and fab1.rication
Introduction
yield. Moreover, process variability is one of the
main problems in current nanometric technologies. This
The
upcoming refers
generations
of integrated inconsistency,
systems, as Many
phenomenon
to unpredictability,
unProcessor
Systems-on-Chip
(MPSoC),
require
drasticfeature
techevenness, and changeability associated with a given
nological
evolutionItsince
they have
terms
or specification.
has became
onereached
of the limits
leadingincauses
offor
power
consumption,
computational
efficiency
and
fabrichip failures and delayed schedules at a sub-micrometric
cation
yield.
Moreover,system
process
variability
is one of
the
scale and
complicates
design
by introducing
uncermain problems in current nanometric technologies. This
phe-nomenon refers to unpredictability, inconsistency,
multaneously
managing both
many voltage
cases,
the energy consumption
of aparameters.
device [4].InSupply
the
only
performance
requirement
is
that
the
tasks
meet
a
can be reduced whenever slack is available in the critical
deadline,
where
a
given
task
has
to
be
computed
before
path, but one has to take care that scaling the voltage ofa a
given
time. Such changes
cases create
to runTherefore,
the promicroprocessor
its opportunities
speed as well.
cessor
at
a
lower
computing
level
and
achieve
the
same
peradapting the supply voltage is very interesting but implies
ceived
performance
while
consuming
less
energy
[5,
6,
7].
the use of Dynamic Frequency Scaling (DFS) to keep
As
a
result,
closed-loop
control
laws
are
required
to
mancorrect the system behavior. The addition of DFS to DVS
age
the energy-performance
tradeoff
in MPSoCs
new
is called
Dynamic Voltage and
Frequency
Scalingand
(DVFS)
strategies
are
developed
in
this
sense
in
this
paper.
and results in simultaneously managing the frequency and
the voltage. In many cases, the only performance
97
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
In addition, embedded integrated systems have two means
of implementation. Firstly, the conventional clocked circuits with their global synchronization – in which one
faces the huge challenge of generating and distributing a
low-skew global clock signal and reducing the clock tree
power consumption of the whole chip – makes them difficult for implementation. Secondly, MPSoCs built with predesigned IP-blocks running at different frequencies need
to integrate all the IP-blocks into a single chip. Therefore, global synchronization tends to be impractical [2].
By removing the globally distributed clock, Globally Asynchronous Locally Synchronous (GALS) circuits provide a
promising solution. A GALS design allows each locally
synchronous island to be set independently, the island becoming a Voltage/Frequency Island (VFI), where an Asynchronous Network-on-Chip (ANoC) is used as the mechanism to communicate between the different VFIs. As a consequence, a GALS-ANoC architecture is more convenient
for DVFS than the standard synchronous approach, because
the power consumption of the whole platform depends on
the supply voltage and the clock frequency applied to each
VFI [4, 8]. A GALS-ANoC architecture also mitigates the
impact of process variability [9], because a globally asynchronous system does not require that the global frequency
was dictated by the longest path delay of the whole chip, i.e.
the critical path. Each clock frequency is only determined
by the slowest path in its VFI.
comes essential in nanometric technologies. The proposed
solution is introduced for a practical microprocessor (i.e.
a MIPS R2000 processor), detailing also the actuators and
sensors. The robust and energy-efficient DVFS control algorithm is depicted in section 3. Stability and robustness
are analyzed too. Fine-grained simulation results are then
presented in section 4. Note that only simulation results
are provided in the present paper. An implementation on a
real chip will be possible after the hardware and/or software
cost of the proposed approach will be evaluated. Nevertheless, preliminary results are obtained for a realistic island
implemented in STMicroelectronics 45nm technology.
2. Controlling uncertainty and handling process variability
The main points of interest of the proposal is to handle
the uncertainty of a processor (or processing node) over a
GALS-ANoC design and reduce its energy consumption.
This is possible by means of automatic control methods.
Using both an ANoC distributed communication scheme
and a GALS approach offer an easy integration of different functional units thanks to a local clock generation [11].
Moreover, this allows better energy savings because each
functional unit can easily have its own independent clock
frequency and voltage. Therefore, a GALS-ANoC architecture appears as a natural enabler for distributed power management systems as well as for local DVFS. It also mitigates
the impact of process variability [9].
An architecture for DVFS and process quality management
is presented in Fig. 1 (note that the description in [12] gives
more details about this architecture and the processing node
circuit.) The operating system (or scheduler) provides a set
of information ref (the required computational speed, in
terms of number of instructions and deadline, for each task
to treat in a given VFI), eventually through the ANoC. This
information about real-time requirements of the application
enables to create a computational workload profile with respect to time. There are also speed sensors (not represented
in Fig. 1) embedded in each processing unit in order to provide real-time measurements of the processor speed ω (in
million of instructions per second for instance). Consequently, combining such a profile with such a monitoring
makes possible to apply a power/energy management allowing application deadlines to be met. On the other hand, the
DVFS hardware part contains voltage and frequency converters, that are a Vdd-hopping and a Programmable SelfTimed Ring (PSTR) for supplying the voltage Vdd and the
frequency fclk respectively. A controller then dynamically
controls these power actuators in order to satisfy the application computational needs with an appropriate management strategy. It processes the error between the measured
speed and the speed setpoint information within a closed-
The present paper is a proof of concept, gathering different techniques together to propose a robust design of an
automatic control algorithm dealing with nano-scale process variations. The setup is based on an energy-efficient
DVFS technique (previously developed in [10]) applied to
a GALS-ANoC MPSoC architecture. A fast predictive control law i) deduces the best two power modes to apply for a
given task and then ii) calculates when to switch from one
power mode to the other, where a power mode is defined
by a voltage/frequency pair supplying a VFI. The switching time instant is such that the energy consumption is minimized while the task fits with its deadline, guaranteeing
good performance. This is repeated for all the tasks to treat.
The control decisions are sent to the voltage and frequency
actuators (respectively a Vdd-hopping and a programmable
self-timed ring) while speed sensors provide real-time measurements of the processor speed. The control strategy is
highly robust to uncertainties since the algorithm does not
need any knowledge on the system parameters. It is dynamically (on line) computed, which ensures that it works for
any type of tasks (whereas only periodic tasks are generally
treated in the literature). Furthermore, the control strategy
is simple enough to limit the overhead it may introduces
(see [10] for further details).
The rest of the document is organized as follows. In section 2, it is explained why a closed-loop architecture be-
98
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
loop system, and by applying a well-suited compensator
sends the desired voltage and frequency code values to the
actuators (denoted Vlevel and flevel ). Consequently, the system is able to locally adapt the voltage and clock frequency
values clock domain by clock domain. Furthermore, the
ANoC is the reliable communication path between the different domains. Data communications between two VFIs
can fix the speed to the slowest communicating node in order to have a secure communication without metastability
problem and an adaptation to process variability too [13].
OS
ref
ANoC
flevel
ref
PSTR
Controller
ω
Vlevel
Vdd
hopping
Vdd
fclk
Processing
node
adjustable clocks. Today, many studies are oriented to SelfTimed Ring (STR) oscillators which present well-suited
characteristics for managing process variability and offering
an appropriate structure to limit the phase noise. Therefore,
they are considered as a promising solution for generating
clocks even in presence of process variability [15]. Moreover, STRs can easily be configured to change their frequency by just controlling their initialization at reset time.
2.3.1 Self-timed rings
A STR is composed of several nested stages which behavior
is mainly based on the “tokens” and “bubbles” propagation
rule. A given stage i contains a bubble (respectively a token) if its output is equal (resp. not equal) to the output
of the stage i + 1. The number of tokens and bubbles are
respectively denoted NT and NB , with NT + NB = N ,
where N is the number of the ring stages. For keeping the
ring oscillating, NT must be an even number. One can think
about this as the duality of designing the inverter ring by odd
number of stages. Each stage of the STR contains either a
token or a bubble. If a token is present in a stage i, it will
propagate to the stage i + 1 if and only if this latter contains a bubble, and the bubble of stage i + 1 will then move
backward to stage i.
ω
Vdd
Voltage/Frequency Island
Figure 1: DVFS control of a GALS-ANoC MPSoC architecture: control of
energy-performance tradeoff in a VFI.
Without lack of generality, the DVFS technique is supposed
to be implemented with two discrete voltage levels Vhigh
and Vlow , with Vhigh > Vlow > 0. Also, ω high and ω low
denote the maximal computational speeds when the system
is running under high and low voltage with its maximal associated clock frequency, with ω high > ω low > 0 by construction.
2.3.2 Programmable self-timed rings
2.1 MIPS R2000 as a processing node
The oscillation frequency in STRs depends on the initialization (number of tokens and bubbles and hence the
corresponding number of stages) [16]. Programmability
can be simply introduced to STRs by controlling the tokens/bubbles ratio and the number of stages. Programmable
Self-Timed Ring (PSTR) uses STR stages based on Muller
gates which have a set/reset control to dynamically insert
tokens and bubbles into each STR stage. Moreover, in order
to be able to change the number of stages, a multiplexer is
placed after each stage [17]. This idea was also extended
to have a fully Programmable/Stoppable Oscillator (PSO)
based on the PSTR. Look-up tables loaded with the initialization token control word (i.e. to control NT /NB ), and the
stage control word (i.e. to control N ) was used to program
the PSTR with a chosen set of frequencies.
The MIPS R2000 is a 32-bit Reduced Instruction Set Computer (RISC). Due to its simplicity in terms of architecture,
programming model and instruction set, as well as availability as an open core, it has been used as a processing node
(see Fig. 1) in our study case. The whole GALS-ANoC island implementation is done using Synopsis Design Vision
tool with STMicroelectronics 45 nm technology libraries
(CMOS045 SC 9HD CORE LS).
2.2 Vdd-hopping for voltage scaling
The Vdd-hopping mechanism was described in [14]. Two
voltages can supply the chip. The system simply goes to
Vlow or Vhigh with a given transition time and dynamics
that depend upon an internal control law (one could refer to
the reference above for more details). Considering that this
inner-loop is extremely fast with respect to the loop considered in this paper, the dynamics of the Vdd-hopping can be
neglected.
2.3.3 PSTR programmability applied to MIPS R2000
Presently, the variability is captured in the design by using
simulation corners, which correspond to the values of process parameters that deviate by a certain ratio from their typical value. In the STMicroelectronics 45 nm CMOS technology, three PVT (Process, Voltage, and Temperature) corners are available, denoted Best, Nominal and Worst. All
standard logic cells were characterized at each of these corners. Since, our main goal is to define the optimal oper-
2.3 Programmable self-timed ring for frequency scaling and
variability management
The application of the proposed DVFS to a system requires
the use of a process variability robust source for generating
99
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
ating clock frequency needed by the processing workload
that compensates for the propagation delay variations due
to the variability impact. Therefore, the critical path delay
of the synthesized MIPS R2000 with respect to the supply
voltage is analyzed at the different PVT corners. The resulting optimal frequencies needed by the MIPS R2000 at two
specified voltage levels are finally defined in Table 1 for the
three process variability corners.
the deadline ∆i . Furthermore, let Λi denote the laxity, that
is the remaining available time to complete the task Ti .
3.1 Speed setpoint building
The presence of deadline and time horizon to compute tasks
naturally leads to predictive control [18]. The main idea of
the predictive strategy is firstly intuitively explained and its
formal expression is given in the sequel.
Table 1: Optimal clock frequencies required to compensate for the process
variability.
3.1.1 Intuitive DVFS control technique
Voltage level
0.95 V
1.1 V
Clock frequency
for different process variability conditions
Worst
Nominal
Best
60 MHz
75 MHz
85 MHz
95 MHz
115 MHz
145 MHz
2.4 Speed sensors for real-time measurement
Any closed-loop scheme requires measurements to compare
the measure with a given setpoint to reach. In the present
study case, speed sensors are embedded in each VFI. They
provide a real performance measurement of the processor
activity, i.e. its computational speed (in number of instructions per second).
Speeds sensors play a critical role and must be carefully
selected. A reference clock fixes the time window where
the number of instructions are counted. Its period is crucial
since it determines the accuracy of the calculated average
speed and the system speed response. Therefore, according
to the set of clock frequencies available for the MIPS R2000
(see Table 1), the reference clock was chosen to be 2 MHz
in order to count a considerable amount of instructions with
a proper system response. To conclude, the computational
speed is now applied in terms of number of instructions executed per 500 ns to the digital controller.
3. Energy-efficient DVFS control with strong
process variability robustness
The control of the energy-performance tradeoff in a voltage
scalable device consists in minimizing the energy consumption (reducing the supply voltage) while ensuring good computational performance (fitting the tasks with their deadline). This is the aim of the controller introduced in Fig. 1,
which dynamically calculates a speed setpoint that the system will have to track. This setpoint is based on i) the
measurement of the computational speed ω (obtained with
the speed sensors) and ii) some information provided by a
higher level device (the operating system or scheduler) for
each task Ti to treat. Information are the computational
workload, i.e. the number of instructions Ωi to execute, and
A naive DVFS technique applies a constant power mode for
each task Ti to treat, as represented in Fig. 2(a). The average speed setpoint, that is the ratio Ωi /∆i , is constant for
a given task. Therefore, if the computational workload of
a given task is higher than the processor capabilities under
low voltage, i.e. Ωi /∆i > ω low , then the VFI executes the
whole task with the penalizing high voltage Vhigh and its
associated frequency in order not to miss its deadline. This
is the case for T2 for instance.
Note that if Ωi /∆i > ω high for a given task to treat, the
execution of the task is not feasible by the desired deadline.
3.1.2 Energy-efficient DVFS control technique
To overcome such intuitive approaches, an energy-efficient
control has been proposed in [10]. The idea is to i) minimize
the energy consumption by reducing as much as possible the
penalizing high voltage level ii) while ensuring the computational performance required so that the tasks meet their
deadline. A task is thus split into two parts, see Fig. 2(b).
Firstly, the VFI begins to run under high voltage Vhigh (if
needed) with the maximal available speed ω high in order to
go faster than required. Then, the task is finished under low
voltage Vlow with a speed lower than or equal to the maximal speed at low voltage ω low which, consequently, highly
reduces the energy consumption, where highly is task dependent. The task is hence executed with a given ratio between high and low voltage. A key point in this strategy is
that the switching time to go from Vhigh to Vlow has to be
suitably calculated in order to meet the task deadline. This
is done thanks to a fast predictive control law.
In fact, the lower is the supply voltage the better will be the
energy savings. For this reason, only one possible frequency
Fhigh is possible when running under Vhigh (in order to
minimize the penalizing high voltage running time). On the
other hand, several frequency levels Flown are possible under low voltage Vlow because, as the energy consumption
could not be reduced anymore, the degree of freedom on
the frequency will allow to fit the task with its deadline (as
much as this is possible). In the present case, two frequency
levels exist under Vlow , with Flow1 ≥ Flow2 . Whereas the
maximal levels Fhigh and Flow1 are determined from the
100
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
speed setpoint ωsp is then directly deduced from the value
of the predicted speed
ω high
if δ(t) > ω low
ωsp (t) =
(2)
ω low
elsewhere
average computational speed setpoint
ω high
Ω2
∆2
ω low
Ω1
∆1
T2
Ω3
∆3
T1
t1
and so are the voltage and frequency levels. Indeed, the system has to run under Vhigh /Fhigh when the required setpoint
is higher than the capabilities under low voltage, else Vlow
will be enough to finish the task. The method for the frequency control decision when the system is running under
low voltage is similar.
The proposed control strategy is dynamically computed.
The online measurement of the computational speed ω ensures that the control law works for any type of tasks, periodic but also non-periodic, because it is not required to
a priori know ω. Moreover, the control algorithm will
react in case of perturbation, e.g. if ω does not behave
as expected, or if the computing workload (Ωi or ∆i ) is
adapted/estimated during the execution of the current task.
On the other hand, the control strategy is simple enough to
limit the overhead it may introduces. Actually, some simplifications are possible for a practical implementation, like
removing the division in (1) (see [10] for further details).
T3
t2
t3
time
t3
time
voltage
Vhigh
Vlow
t1
t2
tVhigh
(a) Intuitive average speed setpoint.
energy-efficient computational speed setpoint
ω high
Ω2
∆2
ω low
Ω1
∆1
T2
Ω3
∆3
T1
t1
T3
t2
k
t3
time
k
t3
time
voltage
Vhigh
Vlow
3.3 Performance and stability
t1
t2
tVhigh
(b) Energy-efficient speed setpoint.
Figure 2: Different computational speed setpoint buildings can be used to save
energy consumption while ensuring good performance.
optimal frequency values in Table 1 (regarding the variability of the chip), the second frequency level at low voltage is
chosen equal to Flow1 /2. This enables to have 3 dB reduction in the power consumption.
3.2 Fast predictive control
A predictive issue can be formulated as an optimization
problem but an optimal criteria is too complex to be implemented in an integrated circuit. Further, a fast predictive control consists in taking advantage of the structure of
the dynamical system to make faster the determination of
the control profile, see e.g. [18]. In the present case, the
closed-loop solution yields an easy algorithm as one simply
needs to calculate the so-called predicted speed δ, defined as
the speed required to fit the task with its deadline regarding
what has already been executed
δ(t) =
Ωi − Ω(t)
Λi
(1)
where Ω(t) is the number of instructions executed from the
beginning of the task Ti . The dynamical energy-efficient
The performance of the proposed control strategy is guaranteed because the execution of a task always starts with the
penalizing high voltage level (by construction of the predictive control law) and the low level will not be applied while
the remaining computational workload is important (higher
than the maximal possible speed at Vlow ). As a result, it is
not possible to make better. Furthermore, even if the voltage/frequency levels discretely vary, the speed setpoint to
track is always higher or equal than required by construction.
On the other hand, the Lyapunov stability is verified. Lyapunov stability is based on an elementary physical constatation: if the total energy of the system tends to continuously
decline, then this system is stable since it is going to an
equilibrium state. Let
V (t) = Ωi − Ω(t)
(3)
be a candidate Lyapunov function. This latter expression
comes from (1) and refers to the remaining workload in the
contractive time horizon of the task. As a result, the Lyapunov function intuitively decreases because the speed of
the processor can only be positive, and so is ensured the stability of a task. This is verified for all tasks to be executed.
3.4 Estimation of the maximal speeds and robustness
The maximal speeds ω high and ω low cannot be directly calculated since they could vary with temperature or location
101
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
(variability), and yet, the control law has to be robust to such
uncertainties. For these reasons, the maximal speeds are estimated. A solution consists in measuring the system speed
for each power mode and using a weighted average of the
measured speed in order to filter the (possible) fluctuations
of the measurement. One can refer to [10] for further details. The proposed estimation of the computational speeds
leads to a control law without any need of knowledge on
the system parameters. This means the strategy is robust
to technological uncertainties since it self-adapts whatever
the performance of the controlled chip. Robustness deals
with the conservation of the stability property when things
do not behave as expected. Unexpected behaviors can be of
two types: a wrong estimation of the number of instructions
to process Ωi or the deadline ∆i , or the presence of process
variability.
In the first case, if Ωi is overestimated (or ∆i is underestimated), the task will be completed before its deadline but
the Lyapunov function V in (3) remains decreasing. On
the other hand, if Ωi is underestimated (or ∆i is overestimated), the deadline will be missed. In this case, the remaining instructions are added to the next task workload in
order to speed up the system. Another solution could be
to use a damping buffer as usually done for video decoding,
see [19, 20] for instance. Note that Ωi and ∆i can also be reestimated during the execution of the task in order to reduce
the error of estimation. Nonetheless, V remains decreasing
during the whole task execution.
3.5 Coarse-grain simulation results
Coarse-grain simulations are performed in Matlab/Simulink
in order to evaluate the efficiency of the proposed controller.
A scenario with three tasks to be executed is run, the number of instructions and deadline for each task being known.
The simulation results are presented in Fig. 3. The top plot
shows the average speed setpoint of each task (for guideline), the predicted speed (for guideline) and the measured
computational speed, while the bottom plot shows the supply voltage. In Fig. 3(a), the system runs during about 80 %
of the simulation time under low voltage. As a result, a reduction of about 30 % and 65 % of the energy consumption
is achieved (in comparison with a system without DVS and
DVFS mechanism respectively).
As the proposed control strategy does not use any knowledge on the system parameters, the controller adapts itself
with these uncertainties. Fig. 3(b) shows how the system behaves in case of 20 % of process variability, that is when the
real performance of the circuit are 20 % less than expected.
One could see that the estimation of the maximal computational speed allows the system to still work, even if the
processing node does not work as expected. Of course, in
order to compensate a lower computational speed induced
by the process variability, the system runs a longer time under the penalizing supply voltage.
More simulations and details are presented in [10].
E = 1.14326 eJ
C = 19302 OPs
Fully discrete control scheme
2 voltage levels and 3 frequency levels −− using of the clock gating
Computational
speeds
[x107 IPS]
Voltage
[V]
In case of process variability, the real computational speed
becomes ωreal (t) = α ω(t), where α ≥ 0 is the unknown
process variability factor, and the system performance is
hence affected. α = 1 means the real process behaves ideally. Otherwise, the performance of the system is weaker
than expected for α < 1 and the system is faster than expected for α > 1. Nonetheless, V in (3) becomes
4
average speed setpoint
predicted speed
measured speed
3
2
1
0
1.1
voltage
0.8
0
0.5
1
1.5
2
2.5
3
3.5
4
Time [x10−6 s]
(a) Simulation results with three frequency levels, one for the high voltage
level and two for the low voltage level.
Robustness to process variabilty (20%)
2 voltage levels and 3 frequency levels −− using of the clock gating
4
and so the stability is still ensured for all α > 0. The convergence rate is reduced for 0 < α < 1 and increased
for α > 1. As a result, the system runs a longer (respectively shorter) time under the penalizing high supply voltage (by construction of the control law) to compensate the
weaker (resp. better) performance. The only unstable case
is for α = 0, which means that the processor does not compute at all. In that case, the operating system (or scheduler) must avoid allocating tasks to this part of the chip.
Also, note that the tasks can meet their deadline as far as
the processor is able to execute the computational workload
in the given time (denoted as feasible tasks), that is while
Ωi /∆i ≤ α ω high .
Computational
speeds
[x107 IPS]
(4)
Voltage
[V]
V (t) = Ωi − αΩ(t)
average speed setpoint
measured speed (0% of variabiity)
measured speed (20% of variabiity)
3
2
1
0
1.1
voltage
0.8
0
0.5
1
1.5
2
2.5
3
3.5
4
Time [x10−6 s]
(b) Simulation results to test the robustness of the controller with 20% of
process variability.
Figure 3: Simulation results of the energy-performance tradeoff control in
Matlab/Simulink.
4. Fine-grain simulation results
In this latter section, fine-grain simulation results are presented for the whole MIPS-R2000 architecture with the
102
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
45 nm CMOS parameter variations. A post layout simulation with Modelsim has been performed with a scenario of
three tasks, the number of instructions and deadline for each
task being known. The simulation results of the system under different process variabilities are shown in Fig. 4. The
different available frequencies for each case were summarized in Table 1.
a) The results under nominal process variability are depicted in Fig. 4(a). Tasks 1 and 3 are completed successfully using the low voltage and frequency levels. For
task 2, the controller speeds up the MIPS R2000 to Vhigh
in order to be able to complete the task at the proposed
deadline. Once the controller has detected that task 2
can be completed with relaxed conditions (that is after
1.52 µs), the system switches back to Vlow .
b) The results under worst process variability are represented in Fig. 4(b). With reduced performance of the
MIPS R2000 (i.e. increased critical path delay), task 2
runs for 3.02 µs under Vhigh , which is twice longer than
processing under nominal process viability.
c) The results when using the best process variability configuration are given in Fig. 4(c). The MIPS R2000 is
able to successfully complete all the three tasks under
Vlow , which adds much more power/energy saving opportunities than under the nominal case. Therefore, our
proposal is able to not only compensate for the delay
variations with different process variability impacts, but
also exploit the enhanced response of the system under
best variability conditions to gain more in terms of energy savings.
To evaluate the proposed DVFS control for GALS-ANoC
architecture, its average dynamic power, energy consumption, area overhead and robustness to process variability are
also analyzed.
4.1 Energy and average dynamic power savings
Under nominal process variability, the energy-efficient
DVFS control (using dynamic set of clock frequencies) is
able to save 21 % of the energy consumption and 51 % of
the average dynamic power consumed by a system without DVFS, while the intuitive average-based strategy (using
fixed set of clock frequencies) only saves 14 and 37 % respectively. Therefore, it appears that the energy-efficient
control is 1.5 more power and energy saving efficient than
the intuitive control (under nominal process variability).
Our proposal has the ability to adapt the set of clock frequencies with respect to the process variability impact.
Thus, the energy-efficient DVFS control was able to exploit
the enhanced performance of the system (i.e. reduced critical path delay) to save more energy consumption (i.e. 25 %
under best process variability impact).
Under worst process variability conditions, the used set of
clock frequencies for a system without DVFS, and even that
for a system with average based DVFS control, violates the
MIPS R2000 critical path delay. As a consequence, the
MIPS R2000 have erroneous output results. Such a processing node has to be neglected and its allocated tasks have
to be reallocated over other high performance processing
nodes. However, with the proposed DVFS control architecture, the MIPS R2000 is still able to complete the allocated
tasks successfully by using the proper set of maximal clock
frequencies. Therefore, this drastically relaxes the fabrication constraints and helps to the yield enhancement.
4.2 Robustness to process variability
As already mentioned, the proposed approach is robust to
technological uncertainties since the control algorithm does
not need any knowledge on the system parameters. Fig. 4
clearly shows how the controller adapts itself with the different (nominal, worst, best) variability corners. The system still works even if the processing node does not work
as expected. However, in order to compensate for a weaker
(respectively stronger) computational speed induced by the
process variability, the system runs a longer (resp. shorter)
time under the penalizing supply voltage level. Of course,
the energy consumption is impacted in consequence.
4.3 Whole energy-efficient control and area overhead
The whole DVFS control system is also evaluated taking
into account the consumption of the actuators and the processing element (see Fig. 1). A GALS-ANoC island is compared with a single processing element (i.e. MIPS R2000)
and with eight controlled elements.
Under nominal process variability, the average dynamic
power and energy saving values of the whole DVFS control system are smaller than but not too far from those presented in previous section (46 and 15 % respectively). Also,
the area overhead due to the proposed control approach is
about 33 %. On the other hand, a single DVFS control system is needed for all the VFIs in a GALS-ANoC system.
Therefore, in a VFI with multiprocessing elements, the efficacy of the DVFS control system is more effective in saving
power/energy consumption (51 and 20 % respectively) and,
moreover, the area overhead of the extra DVFS hardware
is approximately divided by the number of processing elements per a GALS island. For example, the area overhead
in a processing island with eight processors is 4.15 %.
5. Conclusions
In this paper, a survey of different problems facing designers over the nanometric era was first presented. A GALSANoC system was taken as an issue with the application
103
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Figure 4: Timing diagram of the digital controller behavior with 3 different MIPS R2000 workloads under different process variability effects.
of DVFS technique. Also, a closed-loop scheme clearly appeared as necessary in such systems and an architecture was
proposed in this way. The idea to use integrated sensors embedded in each clock domain, as well as a Programmable
Self-Timed Ring (PSTR) oscillator, was presented as one
of the promising solutions to reduce the process variability
impact. A control algorithm has also been detailed, based
on a fast predictive control law. The proposed feedback
controller smartly adapts voltages and frequencies (energy-
performance tradeoff) with strong technological uncertainties. Stability and robustness were analyzed.
A practical validation was realized in simulation for a MIPS
R2000 microprocessor over STMicroelectronics 45 nm
technology to obtain results about the power consumption
and area overhead of each unit in the power management
architecture. Global results for a multicore system were
also presented. Through this example, it was hence demonstrated that a dedicated feedback system associated to the
104
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
GALS-ANoC paradigm and DVFS techniques, with correct
sensors and actuators, is able to achieve better robustness
against process variability. This relaxes the fabrication constraints, thanks to an appropriated strategy, and helps to the
yields enhancement by applying design techniques.
These preliminary results must now be evaluated on a real
test bench. A precise evaluation of the hardware and/or software implementation cost of the proposed control scheme
has also to be performed.
Acknowledgments
This research has been supported by ARAVIS, a Minalogic
project gathering STMicroelectronics with academic partners, namely TIMA and CEA-LETI for micro-electronics,
INRIA/LIG for operating system and INRIA/GIPSA-lab for
control. The aim of the project is to overcome the barrier of
sub-scale technologies (45nm and smaller).
References
[1] B. F. Romanescu, M. E. Bauer, D. J. Sorin, and S Ozev. Reducing the impact of process variability with prefetching and
criticality-based resource allocation. In Proceedings of the
16th International Conference on Parallel Architecture and
Compilation Techniques, 2007.
[2] L. Fesquet and H. Zakaria. Controlling energy and process
variability in system-on-chips: Needs for control theory. In
Proceedings of the 3rd IEEE Multi-conference on Systems
and Control - 18th IEEE International Conference on Control Applications, 2009.
[3] A. P. Chandrakasan and R. W. Brodersen. Minimizing power
consumption in digital CMOS circuits. Proceedings of the
IEEE, 83(4):498–523, 1995.
[4] K. Flautner, D. Flynn, D. Roberts, and D. I. Patel. An energy
efficient SoC with dynamic voltage scaling. In Proceedings
of the Design, Automation and Test in Europe Conference
and Exhibition, 2004.
[5] A. Varma, B. Ganesh, M. Sen, S. R. Choudhury, L. Srinivasan, and J. Bruce. A control-theoretic approach to dynamic
voltage scheduling. In Proceedings of the International Conference on Compilers, Architecture and Synthesis for Embedded Systems, 2003.
[6] T. Ishihara and H. Yasuura. Voltage scheduling problem for
dynamically variable voltage processors. In Proceedings of
the International Sympsonium on Low Power Electronics and
Design, 1998.
[7] J. Pouwelse, K. Langendoen, and H. Sips. Dynamic voltage
scaling on a low-power microprocessor. In Proceedings of
the 7th Annual International Conference on Mobile Computing and Networking, 2001.
[8] P. Choudhary and D. Marculescu. Hardware based frequency/voltage control of voltage frequency island systems.
In Proceedings of the 4th International Conference on Hardware/Software Codesign and System Synthesis, pages 34 –39,
oct. 2006.
[9] D. Marculescu and E. Talpes. Energy awareness and uncertainty in microarchitecture-level design. IEEE Micro, 25:64–
76, 2005.
[10] S. Durand and N. Marchand. Fully discrete control scheme
of the energy-performance tradeoff in embedded electronic
devices. In Proceedings of the 18th World Congress of IFAC,
2011.
[11] M. Krstic, E. Grass, F. K. Gurkaynak, and P. Vivet. Globally
asynchronous, locally synchronous circuits: Overview and
outlook. IEEE Design and Test of Computers, 24:430–441,
2007.
[12] H. Zakaria and L. Fesquet. Designing a process variability robust energy-efficient control for complex SoCs. IEEE
Journal on Emerging and Selected Topics in Circuits and
Systems, 1:160–172, 2011.
[13] T. Villiger, H. Käslin, F. K. Gürkaynak, S. Oetiker, and
W. Fichtner. Self-timed ring for globally-asynchronous
locally-synchronous systems. In Proceedings of the 9th International Symposium on Asynchronous Circuits and Systems, 2003.
[14] C. Albea Sánchez, C. Canudas de Wit, and F. Gordillo. Control and stability analysis for the vdd-hopping mechanism. In
Proceedings of the IEEE Conference on Control and Applications, 2009.
[15] J. Hamon, L. Fesquet, B. Miscopein, and M. Renaudin.
High-level time-accurate model for the design of self-timed
ring oscillators. In Proceedings of the 14th IEEE International Symposium on Asynchronous Circuits and Systems,
2008.
[16] O. Elissati, E. Yahya, L. Fesquet, and S. Rieubon. Oscillation period and power consumption in configurable selftimed ring oscillators. In Joint IEEE North-East Workshop
on Circuits and Systems and TAISA Conference, pages 1 – 4,
2009.
[17] E. Yahya, O. Elissati, H. Zakaria, L. Fesquet, and M. Renaudin. Programmable/stoppable oscillator based on selftimed rings. In Proceedings of the 15th IEEE International
Symposium on Asynchronous Circuits and Systems, 2009.
[18] M. Alamir.
Stabilization of Nonlinear Systems Using
Receding-Horizon Control Schemes: A Parametrized Approach for Fast Systems, volume 339. Lecture Notes in Control and Information Sciences, Springer-Verlag, 2006.
[19] S. Durand, A.M. Alt, D. Simon, and N. Marchand. Energyaware feedback control for a H.264 video decoder. International Journal of Systems Science, Taylor and Francis
online:1–15, 2013.
[20] C. C. Wurst, L. Steffens, W. F. Verhaegh, R. J. Bril, and
C. Hentschel. QoS control strategies for high-quality video
processing. Real Time Systems, 30:7–29, 2005.
105
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Mobile IP Issues and Their Potential Solutions:
An Overview
Aatifah Noureen1, Zunaira Ilyas1,Iram Shahzadi1, Muddesar Iqbal1, Muhammad Shafiq2, Azeem Irshad3
1
Faculty of Computing and Information Technology
University of Gujrat,Gujrat, Pakistan
2
Department of Information and Communication Engineering
Yeungnam University, Gyeongsan, South Korea
3
Department of computer science and software engineering
International Islamic University, Islamabad, Pakistan
{atifanoureen, zunaira.ilyas,}@gmail.com,
[email protected]
[email protected],[email protected]
[email protected]@gmail.com
A typical Internet protocol (IP)could not address the
mobility of nodes and was only designed for fixed
networks where the nodes were improbable to move
from one location to other. An ever-increasing
dependence on network computation makes the use of
portable and handy devices,inevitable. Mobile IP
protocol architecture was developed tomeet such needs
and support mobility. Mobile IPlets the roving nodes to
establish an uninterrupted connection towards internet
without altering the IP address whilemoving to another
network.However, Mobile IP goes through several issues
like ingress filtering, triangle routing, handoff and
Quality of service problems etc. In this paper we discuss
few of those with their possible solutions. That resolves
these problems, reduce the unnecessary load from the
network and enhance the efficiency. Some security
threats on mobile and their solutions are also focused to
secure the communicationduring mobility.
support open standard Protocol suggested by the
Internet Engineering Task Force (IETF) in
November 1996. That makes it possible for the
moving nodes to maintain an un-interrupted
connection towards internet without altering the IP
address while shifting to another network [2]. The
primary reasons that persuading the need of Mobile
IP; making available a scalable, translucent as well
as a safe alternative [3]. In Mobile IP a host node
possesses two addresses at a time, one for the home
network; permanent address and other for the
foreign network which is short-lived and just
legitimate at a particular foreign network [4]. Every
roving node is recognized through its home address
irrespective of its present location over the internet.
When the node is far away to home network it
owed a care-of address to determine its existing
location [5].
Keywords: Mobile IP, Triangle routing, Ingress
1.2 Functional components of mobile IP:
Abstract
filtering, binding updates and COA.
The integral componentsbelong to the mobile
architecture are as follow:
1. Introduction









1.1 Mobile Internet Protocol (Mobile IP)
In standard IP routing, a problem occurs when a
mobile node roams through multiple access
points.It has to reinstate the connection each time
with changing in its IP address due to which the
lively session of the nodes was dropped and
response time of the network is increased [1]. A
novel technology which satisfies the requirement
for smooth connection is Mobile IP; a mobility
Mobile Node (MN)
Home Agent (HA)
Foreign Agent (FA)
Correspondent Node (CN)
Home Agent Address
Care-of-Address (CoA)
Home Network
Foreign Network
Tunneling
106
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
These architectural components are conferred
below:
Foreign network: The network in which the
Mobile Node is currently existent, away from its
Home network.
Mobile Node (MN): Any device which software
facilitates network roaming. A node using the
Mobile IP that moves in different IP subnets. A
permanent (home address) IP address is allotted to
this node which defines the destination of all its
packets. Other nodes use only home IP address to
send packets, regardless of node’s physical
location.
Tunneling: The encapsulating process foran IP
packet in a further IP packet to route it at a place
other than the one defined in the destination field
originally. Prior to promoting it to a suitable router,
when a packet is accepted by home agent, it
encapsulatesthat packet into a new packet, setting
the mobile node’s care-of address in the new
target/destination address field. The path that is
pursued by the new packet is known as the tunnel.
[5]
Home Agent (HA): A router within home network
that delivers packets to the Mobile Node and in
charge to intercept packets whenever the mobile
node is connected to an overseas network. The
home agent is liable for sending these packets to
the mobile node.
1.3 Architecture:
In Figure 1 exemplifies the simple architecture of
mobile IP scenario.
Foreign Agent (FA): A router in foreign network
to which mobile node is connected. A foreign
network router designed to use for Mobile IP. A
mobile node having a ‘foreign agent care-of
address’ all packets are passed on through this
node. The mobile node possibly uses a foreign
agent for registration in the distant network. [6]
Correspondent Node (CN): Any host with which
mobile node communicates. It is a communicating
host with the mobile node. This can be placed on
the foreign network, home network, or any other
place holding potential to send packets to the
mobile node’s home network.
Figure 1: Architecture of Mobile IP
Home Agent Address: IP address of the device
within its home network.
1.4 How mobile IP Works
Whenever a mobile node goes far away from its
home network it listens for advertisement by
FAandtake care-of-address via a foreign agent then
the mobile node records its existing position with
the Foreign Agent and Home Agent (Agent
Discovery and Registration). Packets are delivered
to the home address of the mobile node are
resented by the home agent over the tunnel towards
the care of address in the foreign agent
(Encapsulation). Then the foreign agent passes the
packets to the mobile node on its home address
(De-capsulation). Every time a node shifts to a
network it needs to obtain a new care-of address. A
mobility agent at home network preserves mobility
binding table where as a foreign agent sustains the
visitor list. [1, 8]
Care-of-Address (CoA): IP address of the device
when it is working in a foreign network. The
address used by mobile node for communication
purpose while it is not away from home network.
When the mobile node is using foreign agent’s IP
address as its care-of address it is named as foreign
agent care-of address, and a collocated care-of
address is used when network interface of the
mobile node is provisionally allocated an IP
number on the foreign network [7].
Home Network: The subnet which communicates
to the home address of the mobile node and the
home agent. It is sighted as the “home” spot of
connection for mobile node.
107
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
2.2 Triangle routing problem
Another aspect [11] is that when mobile node is in
foreign network the corresponding node is not
aware of its care of address, it transmits a packet to
the home address of the mobile node which sends it
to the foreign agent which further delivers it to the
mobile node, on the other side the mobile node
communicates directly with the corresponding
node.
Packets are usually sent efficiently both to and
from the mobile node but when corresponding node
sends packet towards mobile node they have to
follow a longer route as compared to optimum.
Since the both transmission directions have used
distinct paths, creates “triangle routing” problem
which decreases network efficiency and places an
unwanted load on the networks [4, 1, 6]. However
Figure 4 illustrates triangle routing problem
scenario in more simple fashion.
Figure 2: Working of Mobile IP
2. Issues in mobile IP:
These are the transitional issues of mobile IPv4 to
Mobile IP v6:
 Ingress Filtering
 Intra-Domain Movement Problem
 Triangular Routing Problems
 Handoff Problem
 Quality of Service Problem
 Fragility problem
2.1 Ingress Filtering problem
Ingress Filtering [9] is an approach carried out by
the firewall enabled systems to ensure that the
particular packets are truly from the network that
they are declaring, if the IP address is not same as
the network these packets are discarded. If the
mobile node transmits packets using a Home
Address rather than Temporary Care of Address
when it is in foreign network then packets are
declined by the program which supports Ingress
Filtering.This problem occurs when both nodes
(sender and receiver) belongs to same home
network [10]. The whole process is shown in
Figure 3.
Figure 4: Triangle Routing
2.3 Handoff Problem
Handoff problem occurswhen mobile node
switches from its current foreign agent to
another[12]. Since, home agent is not aware of the
change of the position of the mobile node so it
tunnels packets to its old foreign agent till it
receives the news registration request from the
mobile node. Thus packets are lost during the hand
off and as a result communication between
corresponding node and mobile node does not
occur successfully [13].
2.4 Intra-Domain Movement Problem
Figure 3: Ingress filtering
108
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
When mobile node changes its position frequently
within a domain and hand off occurs again and
again, a large number of messages are created in
network which decreases the effectiveness of the
network [14].
2.5 Quality of service Problem
Figure 5: Reverse Tunneling
As mobility is the vital ability of telecom wires
networks, besides this phenomenon some other
management issues are also led to be address,
sinceYong Fenghighlighted themas,due to mobility,
unpredictable bandwidth, inadequate network
resources and increasing error rate it is tough to
offer QoS in mobile environment [15].
Reverse tunneling enables the reception of packets
on home network is free from firewall blockage but
it negatively affects the efficiency and safety of the
mobile IP because of inside subnets direct link[17].
3.2 Solutions of Triangle Routingproblem
2.6 Fragility problem
It is simple and easy to configure a single home
agent but it has also a drawback of fragility. As if
the home agent stops working the mobile node also
come to be unapproachable [16].
To tackle the triangle routing problem, many
protocols are planned like forward tunneling and
binding cache, bidirectional route optimization, the
smooth handoff technique, a virtual home agent
and SAIL.
3. ViableSolutions
3.2.1 Route optimization
The most possible solutions of given problems,so
far proposed by the experts, discussed for single
glance overview, are as follow:
Charles Perkins and David Jhosan presented this
protocol in 2000 and it is used in IP Version 6.
Route optimization mobile IP work as it tells the
new care of address to corresponding Nodes.
Nevertheless the communication would be much
more reliable if correspondent node is aware of
care of address of the mobile node and directly
sends the packets directly to the mobile node at
current location without detouring by the home
network [6, 2].Correspondent node can be aware of
the care-of address of the mobile node as the
moving node tells by itself or a home agent
transmits a binding update message to the
correspondent node when forwarding the packet
towards the foreign agent. The correspondent node
then constructs an entry in the binding cache for
further communication[3]. since a mobile node
shifts to another foreign agent, the foreign agent
will maintain a binding cache of the new care-of
address, forward packets towards mobile node
which has moved from its network and also
transmits a Binding Warning message to aware the
home agent about the movement of the node [1].
Handling of triangular routing by route
optimization mobile IP is effective but in a
complex manner. Regular updating is shared by all
of the nodes that are connected [17].
3.1 Solution of Ingress FilteringProblem
For Mobile IP, ingress filtering triggers/creates
massive problems.Ingress filtering limitations may
be improved through an approach named as reverse
tunneling. In Reverse Tunneling foreign agent
receives packet from the mobile node encapsulate it
and transmit it towards home agent, then the home
agent de-capsulate it and further send it to the
corresponding node. [10]In this manner now we
have always topologically truthful addresses
however it probably raises the delay, congestion of
the packets [5].Figure 5 illustrates the whole
process of reverse tunneling.
109
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
corresponding node send messages to the virtual
home agent which further tunnels them towards the
mobile node. This protocol supports transparency
as the corresponding nodes are not aware of
mobility. All the sessions are just for a short period
of time and after that mobile node can change its
virtual home agent which makes sure that every
time mobile node is near to its virtual home agent
[20].
3.2.2 Reverse Routing
Reverse Routing is proposed by Peifang Zhou and
Oliver W. Yang as a substitute to Route
optimization mobile IP. Mobile node deliver packet
to its home network and while moving to foreign
network, it gets care of address. At the next step
mobile node send registration message to the
corresponding node that contain its home network
address as well as care of address. Then the
corresponding node updates its routing table, and
all the further communication from mobile node to
corresponding node takes place according to
updated binding table. It is simpler then Router
optimization as it relay on one message [17].
3.2.5 Surrogate HA
To resolve resolves triangle routing problem idea
of surrogate home agent (SHA) is presented.
Surrogate home agent exist two levels above the
permanent home agent. When a mobile node goes
far away from its home network, it takes care of
address from foreign agent and registers it with its
home agent which further shares it with SHA.
When the corresponding node sends packets they
are tunneled to SHA rather than permanent home
agent than SHA forwards them towards mobile
node current location. As compare to Original
mobile IP this technique decrease the transmission
time, handoff delay, packet loss, and registration
time But it has some limitations as it only provide
better results whenever corresponding node is
above the SHA otherwise not. It is just supportive
in the inter-network. [21]
3.2.3 Location Registers (LR)
In this method the corresponding node first create
link to the database named as home location
register (is a database bridging information of
mobile home IP address and care of address). And
the mobile node always uses home location register
for registration purpose, when the mobile node is
on home network HLR takes responsibility of
identity mapping and when its location changes to
home network, it register its care of address
assigned it by foreign network similar to mobile IP.
When corresponding nodes want to communicate
message at first check the database of Home
location registration and confirm the mobility
binding and deliver data directly to mobile node
care of address. Correspondent node has to update
its binding database before its registration lifetime
expires. Mobile node upholding lists of its all
corresponding nodes that deliver packet during its
existing registration life time and While it shift to
new place it share binding update to entire active
CN’s [18]. It reduces the cost because it does not
require the home address and foreign address but it
create burden on corresponding node in form of
mobility and ways of interaction with home LR
[17].
3.2.6 Other Possible Solutions
Many other protocols has been also proposed so far
to solve the triangle routing problem as forward
tunneling and binding cache [3], bidirectional route
optimization [22], the smooth handoff technique
[20], a virtual home agent [21], and a port address
translation based route optimization scheme [4].
3.3 Solution of Hand off problem
The transmission overall performance during the
mobile node handoff relies on three factors: re
registration, mobile test, and fundamental
interaction with the above layer protocol [13].
Getting local registration via the layered Mobile IP
[23] is a method that helps in decreasing the wait
time for re-registration and enhances the handoff
functionality of Mobile IP. Original FA notification
is also a solution to reduce packets loss by using
the buffer memory [13].
3.2.4 Virtual Home Agent
In this scheme idea of virtual home agent is
presented. Virtual home agent [19] behaves like a
real home agent. Whenever a mobile node moves
to a foreign network it has assigned a virtual home
address against its permanent home address for a
session period. During the session the
110
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
node. It inserts its own CoA in the field of the
spoofed Binding message. When this particular
message is send to corresponding node it will
revise its binding cache and start out
communication. In this way an attacker can able to
see and alter the packets that are for the mobile
node [25].
3.4 Solution of intra domain movement
Problem
Many protocols are designed to reduce problem of
intra domain movement like TeleMIP, Cellular IP,
and HAWAII. They can resolve the problems of
frequently handoff and consequently decreases
packet loss and hand off delay [14].
4.3 Basic Denial of Service Attacks
3.5 Solution of quality of serviceproblem
If the attacker spoofs the binding update and
resends a registration request to a home address
using own CoA. In this manner the attacker will get
all the packets that fail entire connection with
mobile node [14].
To solve end to end Quality of service problem in
mobile IP combination of (DiffServ) differentiate
service and (RSVP) resource reservation protocol
is used [15].
Bombing CoA with unwanted data
Sometimes the attacker places false CoA in the
binding update and reroute the packets to any other
targeted node. The purpose behind that is to bomb
the targeted node with undesired packets. This
could be possible when attacker understands about
the large data stream involving both nodes [26].
This is a very serious attack because targeted node
cannot recognize the attacker and could not do any
think to avoid.
3.6 Solution of Fragility problem
Multiple home agents rather than single one are
used to solve the fragility issues. If one HA break
down there are others who can took the
responsibility to route the data packets towards
mobile node [16].
4. Mobile IP Threats
Mobility leads to a number of different kinds of
threats that might have different effects on the
protection of the protocol. Though MIPv6 have
many advance features against MIPv4. It uses
IPsec protocol to provide security but still it has
some loop holes for security threats. Few of them
are described here.
4.4 Replaying and Blocking Binding
Updates
In this type of attacks the attacker captures the
binding update message and holds the previous
location of the mobile node. When the MN roves to
a new place then it send that binding update to the
corresponding node on the behalf of mobile node
and reroute the packets on the previous spot. In that
manner an attacker can stop binding updates for
MN at its current position till the CN cache entry
exhales [25].
4.1 Spoofing Binding Updates
If binding updates are not attested / authenticated,
an attacker between the mobile node and
corresponding node can spoofs all the binding
update messages. It captures the messages and can
alter the home address/CoA in it and then deliver
them
to
corresponding
node
[24].Thus
communication between mobile node and
corresponding node cannot do successfully and
packets are transmitted to an unwanted node.
4.5 Replay attacks
Sometimes attacker can spoof and store a copy of
true registration request and use it later by
registering false CoA for the MN.The temporal
effects, are also need to be focused to tackle replay
attacks.
4.2 Attacks against secrecy and integrity
4.6 Eavesdropping
Attacker can bust the integrity and privacy between
mobile node and corresponding node as by sending
the false binding message to the corresponding
In this type of attacks an attacker silently listen the
traffic between MN and CN and gather important
111
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
information. Sometimes it just alters the messages
and
sometimes
makes
its
independent
communication to both MN and CN on the behalf
of each other. Both MN and CN’s are unaware of
this and believe that they are communicating
directly on a secure link [27].
5. Possible
Threats
Solutions
binding updates messages. [26] These tests confirm
the home address and care of address efficiently
but may result as the base of many other harasses
[24].
5.2 Solution of Replay attacks
forPotential
To avoid replay attacks the identification field is
produced as it enables the home agents to figure
out about the next value.
These are the possible solutions proposed by the
researchers as countermeasures:
5.3 Solution of Eavesdropping
MIPv6 uses IPsec for security purpose, it
authenticate and encrypt all IP packet during the
communication session. Packets exchanged during
communication are secured using IPSec but there is
no security procedure exists for this kind of attacks
[27].
5.1 Message Authentication
Whenever corresponding node receive the binding
update message, Mobile IP system architecture
shouldauthenticate it to proceed. Following are
some authentication techniques likely to be
considered for this.
6. Conclusion
5.1.1 Use of Public Key Infrastructure (PKI)
Mobility can be considered as amajorinnovationin
communication technologiescontributing valuein
user’s life style. This ability in telecom
environment comes in to existence by the invention
of Mobile IP. Mobile IP is a protocol that permits
mobile system user to shift from one place to
another keeping permanent IP address without
getting the communicating session terminated.
However, this capability also bears technicalissues
and security concerns as presented by the
researchers. This Paper takes into account
thesechallenges besides possible security concerns
faced by the mobile IPtill now. Since the index of
their possible solutions tosecurethe Mobile
IPcommunication architecture and process by
reducingthenetworkoverhead and maximization of
network performance is also discussed for single
glance overview.
Use of public key infrastructure is a typical method
for authentication but in mobile IPv6 it is difficult
to use a single PKI on the internet. [24]
5.1.2 Use of Cryptographically Generated
Addresses
Cryptographically generated addresses are another
method of authentication. In this technique mobile
node’s public key is hashed in the 2nd half of the
HA. The key makes 62-64 bits of IP address which
make it hard for an attacker to discover the key of
the particular address. But there is only 62-64 bits
of address so attacker can find the key by matching
through error and trial methods. [25] Bombing
attacks does not prevent by the Use of
Cryptographically Generated Addresses.
5.1.3 Return Routability Tests
References:
Return routability tests are also used to authenticate
the binding updates and prevent bombing attacks.
Two types of tests are used one for home address
and other for CoA. Whenever the corresponding
node gets a binding update message deliver a secret
key to the home agent which it further tunnels
towards mobile node, on the other hand the
corresponding node also deliver another data
packet with 2nd key at the CoA. The mobile node
utilizes these two keys for the authentication of
[1] Redi, J., and P. Bahl. Mobile ip: (1998). A solution
for
transparent,
seamless
mobile
computer
communications.Fuji-Keizai’s Report on Upcoming
Trends in Mobile Computing and Communications.
[2] Girgis, Moheb R., Tarek M. Mahmoud, Youssef S.
Takroni, and Hassan S. H.(2010). Performance
evaluation of a new route optimization technique for
mobile IP. arXiv preprint arXiv:1004.0771.
112
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
[3] Chandrasekaran,
Janani.
Mobile
ip:
Issues,
[16] Nada,
challenges and solutions. (2009). PhD diss.,
Fayza
A.(2006).On
using
Mobile
IP
Protocols. Journal of Computer Science 2, Volno. 2
Master’s thesis, Department of Electrical and
[17] Mahmood, Sohaib. (2004). Triangle Routing in
Computer Engineering Rutgers University.
Mobile IP. Lahore University of Management
[4] Raj S. B., Daljeet K., and Navpreet K. Mobile
Sciences, Pakistan. http://suraj. lums. edu. pk/-
IP:(2013). A Study, Problems and their Solutions.
cs678s04/2004% 20ProjectslFinals/Group04.pdf.
International Journal of Scientific Research.
[18] Jan, R., Thomas R., Danny Y., Li F. C., Charles G.,
[5] Kumar, Ranjeeth. (2003) MSIP: a protocol for
Michael B., and Mitesh P. (1999). Enhancing
efficient handoffs of real-time multimedia sessions
survivability of mobile Internet access using mobile
in mobile wireless scenarios. PhD diss., Indian
IP with location registers. In INFOCOM'99.
Institute of Technology, Bombay.
Eighteenth Annual Joint Conference of the IEEE
[6] Nada, Fayza. (2007). Performance Analysis of
Computer
Mobile IPv4 and Mobile IPv6. Int. Arab J. Inf.
and
Communications
Societies.
Proceedings. IEEE, vol. 1, pp. 3-11.
Technol. 4, no. 2: 153-160.
[19] Gao, Qiang, and Anthony Acampora. (2000), A
[7] Siozios, Kostas, PavlosE, and AlexandrosK.(2006).
virtual home agent based route optimization for
Design and Implementation of a Secure Mobile IP
mobile IP. In Wireless Communications and
Architecture.
Networking Conference. WCNC. IEEE, vol. 2, pp.
[8] Stoica, Adrian. (2007).A Review on Mobile IP
592-596.
Connectivity and its QoS.International Journal of
[20] Wang, Nen-Chung, and Yi-Jung Wu. (2006). A
Multimedia and Ubiquitous Engineering.
route optimization scheme for mobile IP with IP
[9] Perkins, Charles. (2002). IP mobility support for
header extension. In Proceedings of the 2006
IPv4.IETF RFC 3344,
international
[10] Doja, M. N., and Ravish S. (2012).Analysis of
conference
on
Wireless
communications and mobile computing, pp. 1247-
Token Based Mobile IPv6 and Standard Mobile
1252. ACM.
IPv6 using CPN Tool. International Journal 2,vol
[21] Kumar, C., N. Tyagi, and R. Tripathi. (2005).
no. 7
Performance of Mobile IP with new route
[11] Wu, Chun H., AnnT., Cheng, ShaoT. Lee, JanM.
optimisation
technique.
In
Personal
Wireless
H., and DerT. L. (2002). Bi-directional route
Communications, ICPWC. 2005 IEEE International
optimization in mobile ip over wireless lan. In
Conference on, pp. 522-526. IEEE.
Vehicular
Technology
Conference,
2002.
[22] Zhu, Zhenkai, Ryuji Wakikawa, and Lixia Zhang.
Proceedings. VTC 2002-Fall. IEEE 56th, vol. 2, pp.
(2011). SAIL: A scalable approach for wide-area IP
1168-1172.
mobility. In Computer Communications Workshops
[12] Ayani, Babak. (2002).Smooth handoff in mobile IP.
PhD
diss.,
Master’s
thesis,
Department
(INFOCOM WKSHPS), 2011 IEEE Conference on,
of
pp. 367-372. IEEE.
Microelectronics and Information Technology at
[23] Gustafsson Eva, Jonsson Annika, Perkins Charles
KTH University of California,
E.(2003). IETF Internet Draft. Mobile IPv4
[13] Minhua Y., Yu L.Huimin Z.(2003). Handover
Regional Registration.
Technology in Mobile IP. Telecommunication
[24] Aura, Tuomas. (2004). Mobile IPv6 SecurityIn
Technology.
Security Protocols,. Springer Berlin Heidelberg,pp.
[14] Min S., ZhiminL. (2003). Mobile IP and Its
Improved
Technologies.
215-234.
Telecommunications
[25] Aura, Tuomas, and JariA. (2002).MIPv6 BU attacks
Science
and defenses. Work in Progress
[15] Yong Feng, FangweiLi. (2003).QoS Guarantee over
Mobile IP.
113
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
[26] Arkko, Jari, Vijay D., and Francis D. (2004).Using
He did Master of Science in Computer Science
(MSCS) degree from UIIT, PMAS Arid
Agriculture University Rawalpindi, Pakistan in
2010. He also received the degree of Master in
Information Technology (M.I.T) from The
University of Punjab in 2006. He served as visiting
lecturer in the Federal Urdu University, Islamabad,
Pakistan, for the period of one year since 2009.He
has been serving as Lecturer in Faculty of
Computing
and
Information
Technology,
University of Gujrat, Pakistan since 2010. He has
published more than 10 research papers within the
field of information and communication
technology, at national and international level. His
research interests span the area of QoS issues,
resources and mobility management in mobile adhoc networks.
Azeem Irshad has been doing Ph.D from
International Islamic University, Islamabad, after
completing MS-CS from PMAS Arid Agriculture
University Rawalpindi and is currently serving as
visiting lecturer in AIOU. He has published more
than 10 papers, in national and International
Journals and proceedings. His research interests
include NGN, IMS security, MANET security,
merger of Ad hoc networks with other 4G
technology platforms, multimedia over IP and
resolving the emerging wireless issues.
IPsec to protect mobile IPv6 signaling between
mobile nodes and home agents.
[27] Barbudhe K Vishwajit, Barbudhe K Aumdevi.
“Mobile IPv6:Threats and Solution.” International
journal of application or innovation in engineering
and management. (2013).
Aatifah Noureen is a research scholar under the
enrollment of MS (IT) program in University of
Gujrat, Pakistan since 2013.She enrolled at the
University of Gujrat in 2008 as she realized her
interests in computing and technology and received
her BS degree in Information Technology from
University of Gujrat in 2012.Her research interest
includes networking performance issues in mobile
computing.
Zunaira Ilyasis a student of MS (IT) in University
of Gujrat, Pakistan since 2013. She did BS (IT)
from the same university in 2012. Her research
interest spans the area of mobile ad-hoc networks.
Iram Shahzadi is a research scholar in University
of Gujratsince 2013. She receives her B.Sc and
MSc degrees from University of The Punjab,
Pakistan in 2002and 2008 respectively. She is
serving as a lecturer at Govt. Margazar College for
women, Gujrat since 2009.Her research interests
are QoS of packet radio networks for real
communication.
Dr. Muddesar Iqbal has done Ph.D. from
Kingston University UK in the area of Wireless
Mesh Networks in 2009. He has been serving as
Associate Professor in Faculty of computing and
Information technology, University of Gujrat,
Pakistan, since 2010. He won an Award of
Appreciation from the Association of Business
Executive (ABE) UK for tutoring the prize winner
in Computer Fundamentals module. He also
received Foreign Expert Certificate from State
Administration of Foreign Experts. He is the
authors of numerous papers within the area of
wireless networks. He is also the approved
supervisor for Ph.D. studies of Higher Education
Commission of Pakistan.
Mr. Muhammad Shafiq is the Ph.D. scholar in
department of Information and Communication
Engineering, Yeungnam University, South Korea.
114
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Action recognition system based on human body
tracking with depth images
M. Martínez-Zarzuela 1, F.J. Díaz-Pernas, A. Tejeros-de-Pablos2, D. González-Ortega, M. Antón-Rodríguez
1
University of Valladolid
Valladolid, Spain
[email protected]
2
University of Valladolid
Valladolid, Spain
[email protected]
as binocular and Time-of-Flight (ToF) cameras. Within
the ToF field, different techniques are utilized; the most
commonly used are the extraction of features from depth
images and the processing of a stick figure model (or
skeleton) using depth information (Figure 1).
Abstract
When tracking a human body, action recognition tasks can be
performed to determine what kind of movement the person is
performing. Although a lot of implementations have emerged,
state-of-the-art technology such as depth cameras and intelligent
systems can be used to build a robust system. This paper
describes the process of building a system of this type, from the
construction of the dataset to obtain the tracked motion
information in the front-end, to the pattern classification backend. The tracking process is performed using the Microsoft(R)
Kinect hardware, which allows a reliable way to store the
trajectories of subjects. Then, signal processing techniques are
applied on these trajectories to build action patterns, which feed
a Fuzzy-based Neural Network adapted to this purpose. Two
different tests were conducted using the proposed system.
Recognition among 5 whole body actions executed by 9 humans
achieves 91.1% of success rate, while recognition among 10
actions is done with an accuracy of 81.1%.
Keywords: Body-tracking, action recognition, Kinect depth
sensor, 3D skeleton, joint trajectories.
According to [3] the use of 3D information results
advantageous over the pure image-based methods. The
collected data is more robust to variability issues
including vantage point, scale, and illumination changes.
Additionally, extracting local spatial features from
skeleton sequences provides remarkable advantages: the
way an action is recognized by humans can be
conveniently represented this way, since the effect of
personal features is reduced, isolating the action from the
user who performed it.
1. Introduction
Body tracking and action recognition are study fields that
are nowadays being researched in depth, due to their high
interest in many applications. Many methods have been
proposed whose complexity can significantly depend on
the way the scene is acquired. Apart from the techniques
that use markers attached to the human body, tracking
operations are carried out mainly in two ways, from 2D
information or 3D information [1, 2].
On the one hand, 2D body tracking is presented as the
classic solution; a region of interest is detected within a
2D image and processed. Because of the use of silhouettes,
this method suffers from occlusions [1, 2]. On the other
hand, advanced body tracking and pose estimation is
currently being carried out by means of 3D cameras, such
Fig. 1 Fifteen-joints skeleton model
As stated by [4], one of the main issues existing in action
recognition is related to the extremely large
dimensionality of the data involved in the process (e.g.
noises in 2D videos, number of joints of 3D skeletons…).
115
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
designed in [6]. They manage simultaneous full-body pose
tracking and activity recognition. Another tracking
method is used in [7], where local shape descriptors allow
classifying body parts within a depth image with a human
shape. The results of this point detector show that an
articulated model can improve its efficiency. In [8], it is
presented the construction of a kinematic model of a
human for pose estimation. Key-points are detected from
ToF depth images and mapped into body joints. The use
of constraints and joint retargeting makes it robust to
occlusions. In [9], human poses are tracked using depth
image sequences with a body model as a reference. To
achieve greater accuracy, this method hypothesizes each
result using a local optimization method based on dense
correspondences and detecting key anatomical landmarks,
in parallel. Other approaches to skeleton tracking use
multi-view images obtained from various cameras [10,
11].
This derives in problems such as an increase in
computational complexity and makes more difficult the
extraction of key features of the action. Current
methodologies use four stages to solve this: feature
extraction, feature refinement to improve their
discriminative capability (such as dimension reduction or
feature selection), pattern construction and classification.
Then, by studying the temporal dynamics of the body
performing the movement, we can decode significant
information to discriminate actions. To that extent, a
system able to track parts of the human body for the whole
duration of the action is needed.
However, many action recognition approaches do not
utilize 3D information obtained from consumer depth
cameras to support their system. The existence of state-ofthe-art technology offering positions of the human body in
all three dimensions calls for the emergence of
methodologies that use it to track this information for a
variety of purposes such as action recognition. In addition
to this, an algorithm that is accurate and computationally
efficient is needed.
In [12] challenge of recognizing actions is accounting for
the variability that appears when arbitrary cameras
capture humans performing actions, taking into account
that the human body has 244 degrees of freedom.
According to the authors, variability associated with the
execution of an action can be closely approximated by a
linear combination of action bases in joint spatio-temporal
space. Then, a test employing principal angles between
subspaces is presented to find the membership of an
instance of an action.
This paper is focused on the work field of action
recognition using a 3D skeleton model. This skeleton is
computed by tracking the joints of the human performing
the movement. The emergence of novel consume-depth
cameras, such as Kinect, enables convenient and reliable
extraction of human skeletons from action videos. The
motivation for this paper is to design an accurate and
robust action recognition system using a scheme that
takes into account the trajectory in 3D of the different
parts of the body. As a result, our system is more reliable
than those systems in which only thresholds are taken into
account when determining if a body part has performed a
movement. The movement is classified into a set of
actions by means of a neural network adapted for this
purpose. To that extent, we track the moving body using
the Kinect hardware and OpenNI/NITE software to
extract 3D information of the position of the different
human joints along time.
Apart from ToF cameras, nowadays there exist several
low-cost commodity devices that can capture the user’s
motion, such as the Nintendo Wiimote, Sony Move, or the
Microsoft Kinect camera. This last device allows
perceiving the full-body pose for multiple users without
having any marker attached to the body. The mechanism
used in this camera device is based on depth-sensing
information obtained by analyzing infrared structured
light, to process a 3D image of the environment. It is
capable of capturing depth images with a reasonable
precision and resolution (640x480 px 30fps) at a
commercial cost.
2. 3D skeleton-based approaches
Around this hardware, a community of developers has
emerged and there is a wide variety of tools available to
work with Kinect. A commonly used framework for
creating Kinect applications is OpenNI [13]. OpenNI
software has been developed to be compatible with
commodity depth sensors and, in combination with
PrimeSense’s NITE middleware, is able to automate tasks
for user identifying, feature detection, and basic gesture
recognition. These applications vary from gaming and
rehabilitation interfaces, such as FAAST [14] to
2.1 Action recognition system
The usage of depth cameras has allowed many new
developments in the field of action recognition. In [5],
depth images are processed with a GPU filtering
algorithm, introducing a step further in Real Time motion
capture. A pose estimation system based on the extraction
of depth cues from images captured with a ToF camera is
116
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
actions are: walking, jogging (slow running), gesture
(“hello” and “goodbye”), throw/catch (different styles),
boxing, combo (sequence of walking, jogging and
balancing).
Augmented Reality software. However, the most up-todate Kinect tools are provided in the Microsoft Kinect
SDK [15], released more recently. A previous study to
determine the advantages of Kinect SDK and OpenNI was
carried out in [16]. It was finally concluded that, despite it
lacked some minor joints, OpenNI offers an open
multiplatform solution with almost the same functionality
for Kinect-based applications.
Some datasets, recorded using ToF cameras include depth
information. The ARTTS 3D-TOF database [19] offers
datasets of faces for detection, heads for orientation, and
gestures for recognition. The gestures dataset is composed
by nine different simple actions, such as push and resize,
performed by different users. Other example is the SZU
Depth Pedestrian Dataset [20], which contains depth
videos of people standing and walking for pedestrian
detection. In [6], a database to evaluate a pose estimation
system was proposed, since a synchronized dataset of ToF
captured images was not available online. This dataset
contains ten activities conducted by ten subjects; besides,
each of the movements was recorded 6 times: clapping,
golfing, hurrah (arms up), jumping jack, knee, bends,
picking something up, punching, scratching head, playing
the violin and waving.
In [3], a methodology for working with the Microsoft
Kinect depth sensor is provided. They explain an efficient
way to make use of the information about the human body
(i. e. relationship between skeleton joints). To build their
system, they assume that the distance function based on
the kinetic energy of the motion signals shows enough
discriminative power. Nevertheless, it is a nontrivial
problem to define geometric (semantic) relationships that
are discriminative and robust for human motions.
Taking Kinect skeleton data as inputs, [4] propose an
approach to extract the discriminative patterns for
efficient human action recognition. 3D Skeletons are
converted into histograms, and those are used to
summarize the 3D videos into a sequence of key-frames,
which are labeled as different patterns. Then, each action
sequence is approached as a document of action ”words”,
and the text classification concept is used to rank the
different heterogeneous human actions, e.g. "walking",
"running", "boxing". However, this technique does not
take into account most of the spatial structure of the data.
Ganapathi et al. built a database composed of 28 realworld sequences with annotated ground truth [5]. These
streams of depth images vary from short sequences with
single-limb motions to long sequences such as fast kicks,
swings, self-occlusions and full-body rotations. Actions
can have different length, occlusions, speed, active body
parts and rotation about the vertical axis. Depth
information is captured with a ToF camera, and groundtruth is collected using 3D markers. The scope of this
extensive dataset is to evaluate performance on tracking
systems.
2.2 Datasets
In order to develop a recognition system, it is necessary to
have a set of movements to learn and classify. In this
section, some of the existing datasets that use depth
information are briefly described and a self-made
movement database constructed using a consumer depth
camera is explained. As described in [17], datasets can be
classified according to different criteria such as the
complexity of the actions, type of problem or source.
Heterogeneous gestures are those who represent natural
realistic actions (e.g. jumping, walking, waving…).
Apart from using ToF cameras, another way to obtain 3D
information is by capturing the video via a commercial
Kinect sensor, which offers a better resolution. An RGBD
dataset captured with Kinect is provided by [21],
containing specific daily life actions such as answering
the phone or drinking water. However, to the best of our
knowledge a Kinect dataset which contains heterogeneous
motions such as running, jumping, etc., cannot be found.
Moreover, several other approaches using 3D skeletons,
such as some of the aforementioned, are also evaluated
with this kind of movements.
According to [2], one of the datasets used to evaluate both
2D and 3D pose estimation and tracking algorithms is the
HumanEva database [18]. Their intention is to become a
standard dataset for human pose evaluation. Videos are
captured using a single setup for synchronized video and
ground-truth 3D motion, using high-speed cameras and a
marker-based motion acquisition system. HumanEva-I
dataset consists of four subjects performing six predefined
actions with a certain number of repetitions each. These
With this in mind and taking as a reference a set of
simple actions, we built our own corpus to test the
proposed action recognition system and compare it with
other related approaches. In order to perform complete
action recognition, the desired dataset should contain fullbody natural motions, not based on interfaces or for
specific recognition tasks (dance moves…), performed by
117
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
represent the position of the points that match the main
joints of the human model. With these points, a stick
figure is drawn on the original image, overlaid in the
silhouette of the person.
different people, as in [18]. It consists of ten independent
actions represented by nine different users. As it is shown
in Figure 2, the actions are: bending down, jumping jack,
jumping, jumping in place, running, galloping sideways,
hop-skip, walking, waving one hand and waving both
hands. This database will be further extended to include
more users and action types. The scene is composed of a
cluttered background inside an office, although this
scenario does not need to be maintained.
Fig. 3 An action video is acquired live using a Kinect camera or via the action
dataset, and the human body is tracked from the beginning of the movement to
the end. The obtained 3D skeleton is processed to build motion patterns which
are introduced to the action classifier in order to determine which action is
being performed.
Fig. 2 Example of different movements and users
3.1 Skeleton pre-processing
Once body information is obtained, it is necessary to
break it down into motion data in order to represent an
action. Aggarwal et al. [24] describe a series of steps to
model an action recognition system. To represent the
trajectory of a joint, the coordinate system takes the center
of the image as the origin and then positive and negative
values are assigned to x and y coordinates. Coordinate z is
represented with a positive value, taking the position of
the camera as the origin. The value of these measures is
offered in millimeters.
3. System description
After presenting state-of-the-art 3D-based recognition
systems and datasets, this section describes our approach
derived from this study. Figure 3 depicts the organization
of the proposed system.
In our approach for action recognition, an OpenNI/NITE
wrapper is utilized to access the kinematic skeleton.
OpenNI is an API that can be used to interact with the
information provided by the Kinect technology. One of its
main features includes body tracking of one or more users
in front of the camera, which involves detecting a body
model and offering the coordinates of some of its joints.
OpenNI, in conjunction with NITE [22], supports the
detection and tracking of a kinematic skeleton. As shown
in Figure 1, in the case of Kinect this skeleton consists of
fifteen joints: head, neck, left shoulder, left elbow, left
hand, right shoulder, right elbow, right hand, torso, left
hip, left knee, left foot, right hip, right knee and right foot.
Once the movement starts, the system stores the body
coordinates. In each frame, the system takes from the
coordinate matrix extracted from OpenNI the x-y-z
coordinates of each joint. This coordinate information is
centered with respect to the position of the torso in the
first frame of the movement. This step allows reducing
variability with respect to the initial position and
redundancy of the coordinate values.
3.2 Pattern Generation
As suggested in [3, 23], instead of using edges or regions
within the images for recognition, a coordinate system of
some parts of the human body is utilized. The coordinates
Once an action is tracked, the segmented sequence is
represented into a motion matrix containing the x-y-z
normalized coordinates of the whole movement along
118
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org

p (t )  ( x j , y j , z j ), j  1,15
time, j
. These motions signals
are rectified with respect to the position of the user’s torso
in the first frame, since we consider that the action can be
executed anywhere in the scene. Then, this tracking block
is processed into an action pattern with the purpose to
feed an intelligent system.
~
~
~
~
 X k0 , Re X k1 , Im X k1 ,, Im X k4

9
~
~
~
~
P    Yk0 , Re Yk1 , Im Yk1 ,, Im Yk4 ,

k 1 ~
 Z , Re Z~ , Im Z~ ,, Im Z~
k1
k1
k4
 k0
     , 
      
      
(1)
3.3 Action recognition
After studying several motion representation schemes [25],
it was inferred that using the coordinate points directly in
a time sequence results in enormous vectors when
modeling a motion pattern. In order to transform the raw
trajectory information into a manageable representation,
different compression techniques were analyzed [26]. The
way we reduce dimensionality of joint trajectories is by
applying the Fourier transform. As stated in [27], this
technique is used in many fields, and allows representing
time-varying signals of different duration. Also, by
keeping only the lowest components, the data can be
represented without losing highly relevant information,
smoothing the original signal. This also eliminates
possible existing noise. It is also indicated in [27] that
selecting the first five components of the Fourier
transform, the main information is always preserved.
The core of the system is in charge of carrying out the
action recognition tasks. Henceforth, we will use the term
class to represent each kind of movement. Therefore,
input patterns of the same class represent the same actions
performed by various users or various repetitions of the
same user.
The proposed system uses a neural network capable of
learning and recognizing action patterns P. This neural
network is based on the ART (Adaptive Resonance
Theory) proposed by Grossberg et al. [28], thus includes
interesting characteristics like fuzzy computation and
incremental learning. Due to the nature of Fuzzy logic,
action patterns P need to be normalized altogether,
adjusting their values to the range of [0, 1], before they
can be introduced to the neural network for learning or
testing. Moreover, patterns are complement-coded to enter
the neural network. As a consequence, the action
recognition network works with very large information
patterns of 486 elements.
Unlike in the aforementioned paper, our system not only
follows the trajectory of a single point, but of a whole
skeleton composed by fifteen joints. When building the
motion pattern, the FFT trajectories of each point,

~
~ ~ ~
Pj  FFT ( p j (t ))  ( X j , Y j , Z j )
, are assembled together
to compose a single vector. In order to reduce even more
the dimension of the final motion pattern, only the main
joints are introduced into the final system. It was
determined that the joints to select were head and limbs,
discarding neck, shoulders, hips and torso. This is
because the range of movements of these joints is quite
limited, and also we are subtracting the initial torso
position to represent a relative motion.
The neural network is comprised of an input layer L1, an
output layer L2 and a labeling stage. L1 and L2 are
connected through a filter of adaptive weights wj.
Similarly, L2 is linked to the labelling stage through Ljl
weights. Before neural network training takes place,
weights wj are initialized to 1 and Ljl to zero.
The action patterns P activate L1 using complement
coding I = (P, 1-P). Each input I activates every node j in
L2 with an intensity T j, through adaptive weights wj
(j=0,…, N). Activation is computed using Eq. (2), where
|.| is the L1 norm of the vector, ^ is the fuzzy AND
operator ((p ^ q)i = min(p i, qi)).
Small values for the FFT may add non-relevant
information whereas greater values may eliminate
important information within those first components. A
sub-vector of nine elements is constructed using the real
and imaginary parts of the first five components of the
FFT. The motion pattern is then represented by
assembling the Fourier sub-vectors of each coordinate of
each point, that is 9 FFT elements (0…5) x 3 coordinates
(x,y,z) x 9 joints (k) = 243, as depicted in Eq. (1).
Tj 
I  wj
0.05  w 
(2)
j
In contrast to the ARTMAP neural network [28], a fast
winner tracking node mechanism is included, so that only
those committed nodes in L2 that are linked to
supervision label l would participate in the competition
process. This way, a mismatch between the supervision
label and the label of the label of winning node is avoided.
119
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Also, this mechanism favors a parallel implementation of
the system and speeds-up the training process.
One of the advantages of our system is that it can be
trained to learn new actions and the neural network does
not forget previous trained examples. For this test for
example, it was only necessary to teach the system to
include the new movements that appear in the larger set.
Figure 5 depicts the accuracy rates for this case.
In layer L2 takes place a winner-take-all competition, and
a winner node J so that TJ=max(Tj) is selected. Then, the
resemblance between the action pattern, II, and the
adaptive winner node, wJ is measured using a match
function in Eq. (3)
I I  wJ
II
 threshold
(3)
In case this criterion is not satisfied, a new node is
promoted in L2 and linked through the adaptive weights
Ljl, to the supervision label l. In case the criterion is
satisfied, the network enters in resonance and the input
vector, II, is learnt according to Eq. (4).
wJ 
1
[ wJ  ( I I  wJ )]
2
Fig. 4 Accuracy rates of the Basics test
(4)
Table 1: Confusion matrix of the Basic test
During classification, weights do not change and the
winning node in layer L2 will determine the predicted
label, corresponding to weight Ljl =1.
Basic
Bend
Bend
9
Pjump
Pjump
7
Jump
4. Evaluation
Walk
Two tests were carried out against the prototype of the
action recognition system. For each, different movements
from the dataset were used; the first one, called Basic,
involved the simplest movements (bending down,
jumping, jumping in place, walking and waving one
hand). The evaluation was carried out with the crosscorrelation leave-one-out technique. This involves nine
experiments, each one consisting in learning the actions
of eight users and performing the test with the remaining
subject.
Jump
Walk
2
9
2
7
Wave1
Figure 4 shows the accuracy rates of the set of tests with
the Basic dataset, resulting quite satisfactory. Movements
corresponding to jumping and walking are slightly
confused. The recognition system achieves a classification
success of 91.1%. Table 1 shows the corresponding
confusion matrix, which shows the specific mistakes
made by the system.
Wave1
9
Fig. 5 Accuracy rates of the Complete test
Table 2 presents the confusion matrix for the more
complex test. Although movements such as jumping jack
and waving both hands are successfully introduced into
the system, jumping forward and galloping sideways are
misclassified into the running category. Skipping and
walking actions show also low accuracy in their
recognition results. The average success rate in this case
is 81.1%. These results allow drawing some important
conclusions about the proposed system.
Once the results of this test were obtained, the inclusion
of new movements in the dataset was determined in order
to carry out further experiments. The second set of tests
was carried out including videos for jumping jack,
running, galloping sideways, hop-skip and waving both
hands.
120
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Table 2: Confusion matrix of the Complete test
Complete
Bend
Bend
9
Jack
Pjump
Jack
Pjump
Jump
Run
Side
Skip
Walk
Wave1
Wave2
9
4
Jump
3
2
9
8
1
Side
Run
1
4
4
Skip
2
1
6
Walk
1
1
1
Wave1
6
9
Wave2
9
4.1 Discussion
5. Conclusions
The system is robust recognizing actions that do not
involve very fast transitions and those that involve
different body parts. However, the recognition rate is
decreased when including more actions that do not
follow these premises.
This study presented a body tracking and action
recognition system for heterogeneous actions using
consumer depth cameras, i.e. Kinect. To test the system,
also an action dataset was created from scratch. Since our
system is not only designed to recognize gestures but to
track general actions, the actions that are contained in the
dataset represent whole body movements. The results
obtained show that the proposed system is capable of
performing action recognition for heterogeneous
movements with a high success rate. When the base set of
five actions is used, a mean accuracy of 91.1% is achieved.
By adding another group of similar but more complex
actions, the mean hit rate is 81.1%.
By understanding how the whole system works, it can be
determined the reason of these results. When two joints
follow similar trajectories in the 3D coordinate system,
the components in the action pattern are closer and the
action can be misclassified. This is the reason of the
errors that occur among movements such as walk, run,
skip and jump, when the tracked points share similar
paths. It is also necessary to take into account that
truncating the trajectory signal into the Fourier
dimension eliminates part of the information, smoothing
the differences between similar signals.
Our system is able to track a human body by means of a
commodity depth camera, and software that provides a set
of trajectories. Taking these trajectories as features to
represent an action has been proved to be robust and
reduce the variability that different actors and repetitions
may introduce into the motion signals. Then, the
intelligent network used to classify the actions is capable
of incremental learning, so new actions can be included
into the system once deployed. It is also remarkable that
the processing applied to the motion signals is reversible,
so the original movement can be recovered.
When compared to other approaches that use joint
trajectories, the results are analogous. In [4] the accuracy
ratio is similar, and their confusion matrices also show
misclassification in jumping-waling actions. Authors in
[12] also state the difficulty of distinguishing walking
and running tracked actions using their approach,
lowering the hit rate.
An advantage of our system against other tracking and
recognition approaches is that the original motion signals
are recoverable, since the preprocessing of the motion
patterns can always be inverted. This system does not
need the actions to follow any specific rhythm either.
A possible future work would include the extension of the
dataset, with more varied actions and poses with
trajectories in the third dimension. We will consider the
possibility of opening the dataset for collaboration, so that
new videos could be submitted using an online webpage.
Also, since the system has been designed as a highly
parallelizable algorithm, GPU parallel coding techniques
121
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
[16] Martínez-Zarzuela M, Díaz-Pernas F.J, Tejero de Pablos A,
Perozo-Rondón F, Antón-Rodríguez M, González-Ortega D
(2011) Monitorización del cuerpo humano en 3D mediante
tecnología Kinect. SAAEI XVIII Edition pp. 747-752.
[17] Chaquet J.M, Carmona E.J, Fernández-Caballero A (2013)
A survey of video datasets for human action and activity
recognition. Computer Vision and Image Understanding
117(6): 633-659.
[18] Sigal L, Balan A.O, Black M.J. HumanEva: Synchronized
video and motion capture dataset and baseline algorithm
for evaluation of articulated human motion. International
Journal of Computer Vision 87(1-2): 4-27.
[19] ARTTS
3D-TOF
Database.
Available:
http://www.artts.eu/3d_tof_db.html. Accessed 2013 Apr
21.
[20] Shenzhen University (SZU) Depth Pedestrian Dataset.
Available:
http://yushiqi.cn/research/depthdataset.
Accessed 2013 Apr 21.
[21] Ni B, Wang G, Moulin P (2011) RGBD-HuDaAct: A colordepth video database for human daily activity recognition.
International Conference on Computer Vision Workshops
(ICCV Workshops), 2011 IEEE pp. 1147-1153.
[22] NITE, Algorithmic infrastructure for user identification,
feature detection and gesture recognition. Available:
http://www.primesense.com/solutions/nite-middleware/.
Accessed 2013 Apr 21.
[23] Guo Y, Xu G, Tsuji S (1994) Understanding Human
Motion Patterns. Proceedings of International Conference
on Pattern Recognition Track B: 325-329.
[24] Aggarwal J.K, Park S. Human Motion: Modeling and
Recognition of Actions and Interactions. 3D Data
Processing, Visualization and Transmission pp. 640-647.
[25] Kulic D, Takano W, Nakamura Y (2007) Towards Lifelong
Learning and Organization of Whole Body Motion Patterns.
International Symposium of Robotics Research pp. 113124.
[26] Gu Q, Peng J, Deng Z (2009) Compression of Human
Motion Capture Data Using Motion Pattern Indexing.
Computer Graphics Forum 28(1): 1-12.
[27] Naftel A, Khalid S (2006) Motion Trajectory Learning in
the DFT-Coefficient Feature Space. IEEE International
Conference on Computer Vision Systems pp. 47-54.
[28] Carpenter G, Grossberg S, Markuzon N, Reynolds J, Rosen
D (1992) Fuzzy ARTMAP: A neural network architecture
for incremental supervised learning of analog
multidimensional maps. IEEE Transactions on Neural
Networks 3(5): 698-713.
may be speed-up computational expensive operations,
such as pattern processing and neural network recognition.
References
[1] Poppe R (2007) Vision-based human motion analysis: An
overview. Computer Vision and Image Understanding 108:
4–18.
[2] Weinland D, Ronfard R, Boyer E (2011) A survey of
vision-based
methods
for
action
representation,
segmentation and recognition. Computer Vision and Image
Understanding 115: 224–241.
[3] Raptis H.H.M, Kirovski D (2011) Real-Time Classification
of Dance Gestures from Skeleton Animation. Proceedings
of the 2011 ACM SIGGRAPH, ACM Press pp. 147–156.
[4] Chen H.L.F, Kotani K (2012) Extraction of Discriminative
Patterns from Skeleton Sequences for Human Action
Recognition. RIVF International Conference, IEEE 2012
pp. 1–6.
[5] Ganapathi V, Plagemann C, Koller D, Thrun S (2010) Real
Time Motion Capture Using a Single Time-Of-Flight
Camera. Computer Vision and Pattern Recognition pp.
755-762.
[6] Schwarz L.A, Mateus D, Castañeda V, Navab N (2010)
Manifold Learning for ToF-based Human Body Tracking
and Activity Recognition. Proceedings of the British
Machine Vision Conference 80: 1-11.
[7] Plagemann C, Ganapathi V, Koller D, Thrun S (2010)
Real-time identification and Localization of Body Parts
from Depth Images. IEEE International Conference on
Robotics and Automation pp. 3108-3113.
[8] Zhu Y, Dariush B, Fujimura K. Kinematic self retargeting:
A framework for human pose estimation. Computer Vision
and Image Understanding 114: 1362-1375.
[9] Zhu Y, Fujimura K (2010) A Bayesian Framework for
Human Body Pose Tracking from Depth Image Sequences.
Sensors 10: 5280-5293.
[10] Yániz C, Rocha J, Perales F (1998) 3D Part Recognition
Method for Human Motion Analysis. Proceedings of the
International Workshop on Modelling and Motion Capture
Techniques for Virtual Environments.
[11] Chen D, Chou P, Fookes C, Sridharan S (2008) MultiView Human Pose Estimation using Modified Five-point
Skeleton Model.
[12] Sheikh M.S.Y, Sheikh M (2005) Exploring the Space of a
Human Action. International Conference on Computer
Vision 2005, IEEE 1: 144–149.
[13] OpenNI, Plataform to promote interoperability among
devices, applications and Natural Interaction (NI)
middleware. Available: http://www.openni.org. Accessed
2013 Apr 21.
[14] Suma E.A, Lange B, Rizzo A, Krum D.M, Bolas M (2011)
FAAST: The Flexible Action and Articulated Skeleton
Toolkit. Virtual Reality Conference pp. 247-248.
[15] Microsoft Kinect SDK, Kinect for Windows SDK.
Available:
http://www.microsoft.com/enus/kinectforwindows/. Accessed 2013 Apr 21.
Mario Martínez Zarzuela was born in Valladolid, Spain; in1979. He
received the M.S. and Ph.D. degrees in telecommunication
engineering from the University of Valladolid, Spain, in 2004 and 2009,
respectively. Since 2005 he has been an assistant professor in the
School of Telecommunication Engineering and a researcher in the
Imaging & Telematics Group of the Department of Signal Theory,
Communications and Telematics Engineering. His research interests
include parallel processing on GPUs, computer vision, artificial
intelligence, augmented and virtual reality and natural humancomputer interfaces.
Francisco Javier Díaz Pernas was born in Burgos, Spain, in1962.
He received the Ph.D. degree in industrial engineering from Valladolid
122
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
University, Valladolid, Spain, in 1993. From 1988 to 1995, he joined
the Department of System Engineering and Automatics, Valladolid
University, Spain, where he has worked in artificial vision systems for
industry applications as quality control for manufacturing. Since1996,
he has been a professor in the School of Telecommunication
Engineering and a Senior Researcher in Imaging & Telematics Group
of the Department of Signal Theory, Communications, and
Telematics Engineering. His main research interests are applications
on the Web, intelligent transportation system, and neural networks for
artificial vision.
Antonio Tejero de Pablos was born in Valladolid, Spain, in 1987.
He received his M.S. in Telecommunication Engineering and M.S. in
ICT Research from the University of Valladolid in 2012 and 2013
respectively. During his academic career he has collaborated on
various projects with the Imaging & Telematics Group of the
Department of Signal Theory, Communications and Telematics
Engineering. Now he works as a trainee researcher at NTT
Communications Science Lab, in Japan. His research fields of
interest are Action Tracking & Recognition, Augmented Reality based
interfaces and General Purpose GPU Programming.
David González Ortega was born in Burgos, Spain, in 1972. He
received his M.S. and Ph.D. degrees in telecommunication
engineering from the University of Valladolid, Spain, in 2002 and 2009,
respectively. Since 2003 he has been a researcher in the Imaging and
Telematics Group of the Department of Signal Theory,
Communications and Telematics Engineering. Since 2005, he has
been an assistant professor in the School of Telecommunication
Engineering, University of Valladolid. His research interests include
computer vision, image analysis, pattern recognition, neural networks
and real-time applications.
Míriam Antón Rodríguez was born in Zamora, Spain, in 1976. She
received her M.S. and Ph.D. degrees in telecommunication
engineering from the University of Valladolid, Spain, in 2003 and 2008,
respectively. Since 2004, she is an assistant professor in the School
of Telecommunication Engineering and a researcher in the Imaging &
Telematics Group of the Department of Signal Theory,
Communications and Telematics Engineering. Her teaching and
research interests include applications on the Web and mobile apps,
bio-inspired algorithms for data mining, and neural networks for
artificial vision.
123
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Contention-aware virtual channel assignment in
application specific Networks-on-chip
Mona Soleymani1 , Fatemeh Safinia2 , Somayyeh Jafarali Jassbi3 and Midia Reshadi4
1,2
3,4
Department of Computer Engineering Science and Research branch, Islamic Azad
University, Tehran, Iran
1
[email protected]
2
[email protected]
Faculty Member of Computer Department, Science and Research Branch, Islamic Azad University, Tehran, Iran
3
[email protected]
4
[email protected]
Abstract
a significant challenge[4],[5],[8].Different application
specific NoCs have specified traffic pattern[13].
Occasionally, the data flow control of multiple paths may
overlapped in these traffic patterns. When two or more
packet contend for one path, congestion appears and as
said, consequently the network influenced drastically by
delay.
Nowadays lots of processing elements should be placed on a
single chip because of ever complexity of applications. In order
to connect this number of processing elements, providing a
interconnection infrastructure is essential. Network on Chip
(NoC) is used as an efficient architecture for chip
multiprocessors, which brings a reasonable performance and
high scalability. A relevant challenge in NoC architectures is
latency that increased in the presence of network contention. In
this paper, we utilized virtual channels in buffer structure of
routers. We calculated end to end latency and link utilization in
NoCs by considering different numbers of virtual channels
ranging from 2 to 5. The simulation results show that using
virtual channels for specified routers in 2D mesh 4x4 decreases
end to end latency up to 17% . It also improves link utilization
and throughput significantly.
Keywords: Network on Chip, virtual channel, path-based
contention, latency.
This discussion leads us to focus on delay as major
challenges in data transmitting of specific-application
NoCs[15]. In this article we proposed a solution by the
help of virtual channels to reduce delay.
Generally, in NoC architectures, the global wires are
replaced by a network of shared links and multiple
routers transmitting data flows through the cores. As
said, one of the main problems in these architectures is
traffic congestion that should be avoided in order to
optimize the efficiency and performance of NoCs.
1. Introduction
In addition, the contention that created in some special
simultaneous data flows through the cores and path also
decreases the network performance, especially latency
and throughput. We should consider that it is very hard
to eliminate the contention completely in such networks
but there are some solutions to control these contentions
and degrade the bad impact of it on the whole data flows
and performance in the network[16].
With technology advances, the number of cores
integrated into a single chip is increasing[1]. Traditional
infrastructures such as non-scalable buses cannot handle
the complexity of inter-chip communications. NoC
(Network on Chip) architectures were appeared as a new
paradigm of SoC design offering scalability, reusability
and higher performance [1],[2]. NoC is a set of
interconnected switches, with IP-cores connected to
these switches.
In this work by considering a special core graph as a
sample, we use our proposed technique foe specified
cores that contention would happen in them.In the
technique proposed in this paper for solving contention
in NoCs, virtual cannels are used for controlling pathbased contentions.
Nowadays, one of the principal research issues that onchip interconnection network designers Encountered is
the network’s performance and efficiency [6]. In other
words, when the network becomes congested some
problems such as contention which would be have a
direct impact on NoC’s delay appeared that should be
solved. Virtual channels are a solution for contentions
but the way and the number of them that should be use is
The structure of this article is organized as follows:
Section 2 reviews the fundamental structure of NoC,
preliminaries are presented in Section 3. Following that,
124
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
in Section 4, path-contention issue is described by
motivational example and an efficient allocation virtual
channels are proposed for some routers in section 5.
Experimental results are given and discussed in
Section 6. Finally, concludes the paper.
single buffer for each port in the router structure, one
data flow behind of another should be waited until the
area of buffer on related port to becoming empty.
Virtual channels originally introduced to solve the
deadlock problem, although and therefore they can be
used to improve network latency and throughput[8].
Assigning multiple virtual paths to the same physical
channel is the nature of VC flow control. General
structure of virtual channel is presented in figure2. As
shown each input port has a physical buffer that divided
into two virtual channels, it means that instead of one
channel which can be placed by four flits we can use two
separated channels that each of them can be managed
two flits alone.
2. Fundamental structure of NoC
NoC architecture consists of three components : routers,
links and processing elements[14]. as shown in Figure 1
each router is composed of five input and five output
ports. One pair of input/output ports is connecting to
processing element while the other four pairs are
connected to four adjacent routers in four directions
(north, south, east and west). Every input port has a FIFO
buffer and link controller, The purpose of buffers is to
act as a holding area, enabling the router to manipulate
data before transmitting it to another router and LC
controls the empty area in the buffer for coming a new
packet/flit from neighbor routers. The transferring of
data between the input and output ports of the switch is
carried out by the crossbar, whose active connections are
decided by the arbiter which is a circuit used to make
decisions which controls the crossbar.
Fig.2. virtual channel
3. PRELIMINARIES
In this section, we introduce some related concepts first.
The network contention that occurs into network on chip
communications can be classified into three types:
source-based contention, destination-based contention
and path-based contention[3].
The source-based contention (figure.3(a)) is appeared
when two variety traffic flows are distributed from the
same source on the same link .The destination-based
contention (figure.3(b)) occurs when two traffic flows
have the same destination, in this case it is common to
see a contention on the link which two traffic flows are
leading to the same destination at the end of the
considered link. the path-based contention (figure.3(c))
occurs when two data flows which come from the
different sources, and go towards the
different
destinations contend for the same links somewhere in the
network, (in other words some links that are common
between two traffic flows are prone to contention.)
Fig.1. structure of tile (router and processing element)
In this paper, we consider a 2-D mesh as our topology
with XY routing where the packet is first routed on the X
direction and then on the Y direction before reaching its
destination [7].
Wormhole switching is a well-known technique of
multiprocessor network[4]. The structure of buffers in
the wormhole switching is based on flit units that allows
minimization of the size of buffers; In other words,
power consumption reducing is caused by flit
segmentation instead of packet segmentation.
In the source-based contention, time scheduling method
for transmission of data can be a good solution for
avoidance of contention inside of specified source core.
Similarly, in the destination-based contention, scheduling
Deadlock is a critical problem in wormhole switching
that occurs when one router used by multiple data flows
which each of them has different destination. By using a
125
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
packets to send to the same destination can has
significant impact on reducing this kind of contention.
figure.4(b) which represent a network of switches (cores
are ignored for simplicity) and several paths onto links
are shown.
Unlike two previous methods path-based contention are
not improved by timing technique, so another methods
are required for optimizing this contention.
We apply XY routing and wormhole switching for data
transmission with 4 flit per packet. As seen in figure.4(c)
there are content in some paths, for instance five paths
have a common link across transmitting their data.
Network contention issue affects performance, latency
and communication energy consumption.
Each specific path that is under XY routing in the traffic
pattern shown in Figure.4(c) , composed of some links
that explain as bellow:
For better explaining the technique proposed in this
paper, we firstly describe the following definitions:
Definition: The core graph is a directed graph G(V,P).
Each vertex vi ϵ V indicates a core in communication
architecture. A directional path pi,j ϵ P illustrates a
communication path for data flow from core-source vi to
the core-destination vj. The weight of pi,j that is shown
by W(pi,j) represented the data flow volume (bits) from
core vi to core vj. .
P0,15 = l0,1 + l1,2 + l2,3 + l3,7 + l7,11 + l11,15
P1,11 = l1,2 + l2,3 + l3,7 + l7,11
P7,15 = l7,11 + l11,15
P6,15 = l6,7 + l7,11 + l11,15
P5,11 = l5,6 + l6,7 + l7,11
Notice that Path-contention occurs on the paths that have
overlapped on the same links. When pi,j ∩ pp,q ≠ Ø, then
there would be at least one path contention. we calculate
the content of α for each link that has path-contention in
this traffic pattern.
α(l1,2) = 2, α(l2,3) = 3, α(l3,7) = 2, α(l7,11) = 5,
α(l11,15) = 3, α(l6,7) = 2
(a)
(b)
(c)
Node u5 to node u15 wants to send a packet based on
XY routing, it goes forward two hops to east (first hop to
u6 and second hop to u7) and then goes down as two
hops.
Fig.3. (a) source-based contention (b) destination-based contention (c)
path-based contention
Similarly, node u1 to sent a packet that allocated to u15
destination should goes east to delivery to u3, then goes
down as three hops. Like two previous data flows, the
packets belong to node u5, u0 and u7 are delivered to
destinations across some common links.
Every specific path from core vi to core vj consists of
some links that demonstrated by Lm,n which means a
directed link from cm to vn ( it is obvious that vm , vn are
always considered as adjacent cores). As said in
section… when two or more paths contend at the same
time for the same link for transferring their data flow,
path-contention occurs. α(Li,j) is a parameter used for a
link from core vi to core vj that shows how many path
contends simultaneously on accessing to it.
As described in the mentioned scenario five flows of data
can be inferred some paths, in more detailed some links,
are common, hence these can be prone to content. In this
example, the link between u7 and u11 has five different
data flows, as shown in figure.4(c) although It may seen
that some of these contention can be appeared as a
destination-based contention, we address them to pathbased contention in this paper.
4. Motivational example
As a starting point we assume to have an application that
be mapped onto a 2-D mesh topology.
The ratio of path-based to source-based contention and
the ratio of path-based to destination-based contention
increase linearly with the network size [3]. therefore in
this paper we focus to reduce path-based contention
since this has the most significant impact on the packet
latency and can be mitigated through the mapping
process. hence to overcome this problem we proposed
using virtual channels in this paper.
the communication between the cores of the NoC are
exposed by the core graph[9]. The connectivity and link
bandwidth are presented by the NoC graph. As shown in
figure.4 general application such as figure.4(a) can be
mapped (as with any mapping algorithm) onto a 2-D
mesh topology with dimension of 4×4, to show the
impact of path-based contention consider the example of
126
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
500
v4
u1
u2
u3
u0
u1
u2
u3
u4
u5
u6
u7
u4
u5
u6
u7
u8
u9
u10
u11
u8
u9
u10
u11
u12
u13
u14
u15
u12
u13
u14
u15
v6
00
16
91
0
v0
u0
v9
v14
v15
31
5
91
51
2
v1
v3
0
v5
v10
v12
0
72
00
18
v11
v13
v8
v7
v2
(a)
(b)
(c)
Fig.4. (a) core graph (b) NoC graph (c) mapping
the u0 and u1 flits go through its destination by another
virtual channel and u7 processing element has a single
virtual channel separately.
5. Proposed Technique
In this section we focus on buffer structure as a
fundamental element of router to reduce contentions.
using a single buffer into router structure does not aid
the contention problem remarkably. Take the case of
with no virtual channels, flits 0 and 1 will prevent flits 5,
6 and 7 from advancing until the transmission of flits 0
and 1 have been completed (figure.5).
(a)
In this paper, we consider 5 flits as a buffer sizing for
any port of each router. In the virtual channel allocation
how many flits we should allocate to each virtual channel
Based on this assumption there are 5 flits that should be
divided between virtual channels proportionally. As an
illustration, u1 has two virtual channels on output east
port: one virtual channel is allocated to u0 and another
belongs to u1 processing element. In this case
considering to the volume of data can determine the
number of flits in each virtual channel. The data volume
of u0 is 910Mb and the data volume of u1 is 512Mb
(figure.4(a)). Hence, u0 gives 3 flits because of more
data volume and 2 remaining flits are allocated to u1.
(b)
Fig.5. (a) router7 without VC (b) router7 with three VCs
In this paper, we propose the use of virtual channels for
some routers that are placed on the path contention. The
number of virtual channels for any port of each router
can be modified. Increasing the number of virtual
channels has a direct impact on router performance,
through their effect on the achievable hardware cycle
time of the router[4]. For example, in u7 we can define
three VCs for south port, one VC for flits coming from
north port (u0, u1),one VC for flits coming west port (u5,
u6) and one single buffer( one VC) for processing
element port.
The paths u0àu15, u1àu11, u7àu15, u5àu11 and
u6àu15 have common links, implicitly there are
contention pending these paths. however, the flits of
u5 and u6 from one virtual channel and on the other hand
Fig.6. VCs for some routers on path-contention
127
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
6. Experimental Results
link utilization
In this paper we have used HNoC which is developed in
Omnet++ environment [18] in order to evaluate the
functionality of Networks-on-Chip[10],[11],[12]. This
simulator makes it possible to assess different traffic
patterns with different packet generation rate. However,
We define 4×4 mesh and 4B as flit sizing.
1
We first evaluate our network from the path-based
contention point of view and routers that encounter with
this kind of contention are selected. After that, virtual
channels are allocated to specified ports of these routers
that path-based contention occurred in them. For the
various numbers of VCs ranges from 2 to 5, different
results are achieved from simulation.
number of recieved packets
end to end latency
600
400
200
0
4
5
number of VCs
)×(
)×(
1
2
3
4
5
Table.1. throughput for different number of VCs
Number of VCS
1
2
3
4
5
Throughput can be defined in a variety of different ways
depending on the specifics of the implementation[17].
We used the definition of throughput based on[17] where
for massage passing system, TP can be calculated, as
follows in Eq.(1):
(
4000
3500
3000
2500
2000
1500
1000
500
0
Fig.9. average number received packets results for different
number of VCs
By modifying the number of VCs the link utilization
differs. Figure8 show that using 5 VCs has the highest
link utilization in compared with 2, 3 and 4 VCs. More
virtual channels lead to more virtual independent path
between routers hence, the bandwidth of links are
utilized efficiently.
(
5
number of VCs
Fig.7. average end to end latency results for different number of VCs
=
4
of IP blocks has been 16 and the total time is 100 µs.
Also. according to proposed technique in this paper , the
number of virtual channels varies from 2 to 5. The TP
results are obtained from equation 1 can be seen in table
1. In this table, we observe that throughput(TP) is
slightly higher when using 5 VCs compared to the case
of without VC.
800
3
3
Fig.8. average link utilization results for different number of VCs
1000
2
2
number of VCs
Simulations are conducted to evaluate the end to end
latency, link utilization and throughput. However, the
most important achievement is the fact that the end to
end latency is reduced to 17% by increasing the number
of virtual channels as seen in figure7.
1
70
60
50
40
30
20
10
0
)
ℎ)
Throughput
0.0032022
0.0035555
0.0038215
0.0040455
0.0041890
7. Conclusions
In this paper we have explored the influence of virtual
channels in the presence of path contention. In
application-specific NoC that mapped on any topology ,
path-contention is appeared as a problem. we show that
selecting the appropriate number of virtual channels
based on data volume for some specified routers can
(1)
As shown in figure9 we calculated the number of
received packets which are equal to (total message) ×
(message length). In the simulation process, the number
128
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
significantly improve properties of network contention.
Simulation has been carried out with a number of
different virtual channels. the experimental results show
that by using our proposed method an improvement is
obtained in end to end latency, link utilization and
throughput parameters.
Proceeding of Design, Automation and Test in Europe
Conference and Exhibition, 2, 896.
[14] Shiuan Peh, L., & Jerger, N. E. (2009). On-Chip Networks
(Synthesis Lectures on Computer Architecture). Morgan
and Claypool publisher, Mark D. Hill Series Editor,
Edition1.
[15] Seiculescu, C., Rahmati, D., Murali, S., Benini, L., De
Micheli, G., & Sarbazi-Azad, H. (2013). Designing Best
Effort Networks-on-Chip to Meet Hard Latency
Constraints. The Journal of ACM Transaction on
Embedded Computing Systems (TECS), 12(4), 109.
References
[1] Benini, L., & De Micheli, G. (2002). Network on Chip: A
New SoC Paradigm. IEEE Trans. Computer, 35(1), 70.
[16] Lin, S. Y., Huang, C. H., & Chao, C. H. (2008). TrafficBalanced Routing Algorithm for Irregular Mesh-Based
On-Chip Networks. The Journal of IEEE Transaction on
Computers, 57(9).
[2] Daly, W., & Towles, B. (2001). Route Packets, Not Wires:
On-Chip Interconnection Networks. Proceeding of Design
Automation Conference (DAC), 684.
[3] Chou, C. L., & Marculescu, R. (2008). Contention-Aware
Application Mapping for Network on Chip Communication
Architecture. IEEE International Conference, 164.
[17] Pande, P. P., Grecu, C., Jones, M., & Ivanov, A. (2005).
Performance Evaluation and Design Trade-Offs for
Network-on-Chip Interconnect Architectures. The
Journal of IEEE Transaction on Computers, 54(8), 1025.
[4] Duato, J., Yalamanchili, S., & Lionel, N. (2002).
Interconnection Networks. Morgan Kaufmann.
[18] Varga, A. (2012). The OMNeT++ discrete event
simulation system.
Proceedings of the European
Simulation Multi conference (ESM’2001), 319.
[5] Dally, W. J., & Towles, B. (2003). Principles and Practice
of Interconnection Networks. The Morgan Kaufmann
Series in Computer Architecture and Design.
Mona Soleymani received her B.Sc. from Tehran central branch
of Islamic Azad University, Tehran, Iran in 2010 in Hardware
engineering and her M.S. from science and Research Branch of
Islamic Azad University, Tehran, Iran in 2013 in Computer
architecture engineering. She is currently working toward the
Ph.D. in computer architecture engineering at the Science and
Research Branch of IAU. Her research interests lie in Network
on Chip and routing, switching and mapping issues in on-chip
communication networks.
[6] Bjerregaard, T., & Mahadevan, S. (2006). A Survey of
Research and Practices of Network-on-Chip. Journal of
ACM Computing surveys (CSUR), 38(1).
[7] Kakoee, M. R., Beatacco, V., & Benini, L. (2011). A
Distributed and Topology-Agnostic Approach for On-line
NoC Testing. Networks on Chip (NoCS), Fifth IEEE/ACM
International Symposium, 113.
[8] Mullins, R., West, A., & moore, S. (2012). Low-Latency
Virtual-Channel Routers for On-Chip Networks.
Proceedings of the 31st annual international symposium on
Computer architecture, 188.
Fatemeh Safinia received her B.Sc. from Tehran central branch
of Islamic Azad University,Tehran, Iran in 2010 in Hardware
engineering and her M.S. from science and Research Branch of
Islamic Azad University, Tehran, Iran in 2013 in Computer
architecture engineering. She is currently working toward the
Ph.D. in computer architecture engineering at the Science and
Research Branch of IAU. Her research interests lie in Network
on Chip and routing, switching and mapping issues in on-chip
communication networks.
[9] Murali, S., & Mecheli, G. De. (2004). BandwidthConstrained Mapping of Cores onto NoC Architectures.
Design, Automation and Test in Europe Conference and
Exhibition, 896.
Somayyeh Jafarali Jassbi received her M.Sc. degree in
computer architecture from Science and Research Branch of
Islamic Azad University(SRBIAU), Tehran, Iran in 2007. She also
received her Ph.D. degree in computer architecture from
SRBIAU, Tehran, Iran in 2010. She is currently Assistant
Professor in Faculty of Electrical and Computer Engineering of
SRBIAU. Her research interests include Information
Technology(IT), computer arithmetic, cryptography and network
security.
[10] Walter, I., Cidon, I., Ginosar, R., & Kolodny, A. (2007).
Access Regulation to Hot-Modules in Wormhole NoCs.
Proceedings of the First International Symposium on
Networks-on-Chip, 137.
[11] Ben-Itzhak, Y., Zahavi, E., Cidon, I., & Kolodny, A.
(2013). HNOCS: Modular Open-Source Simulator for
Heterogeneous NoCs. International Conference on
Embeded Computer Systems(SAMOS), 51.
Midia Reshadi received his M.Sc. degree in computer
architecture from Science and Research Branch of Islamic Azad
University (SRBIAU), Tehran, Iran in 2005. He also received his
Ph.D. degree in computer architecture from SRBIAU, Tehran,
Iran in 2010. He is currently Assistant Professor in Faculty of
Electrical and Computer Engineering of SRBIAU. His research
interests include Photonic NoCs, fault and yield issues in NoCs,
routing and switching in on-chip communication networks. He is
a member of IEEE.
[12] Ben-Itzhak, Y., Cidon, I., & Kolodny, A. (2012).
Optimizing Heterogeneous NoC Design. Proceeding of
the International Workshop on System Level Interconnect
Prediction, 32.
[13] Murali, S., & De Micheli, G. (2004). BandwidthConstraint Mapping of Cores onto NoC Architectures.
129
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Optimization Based Blended Learning Framework for
Constrained Bandwidth Environment
Nazir Ahmad Suhail1, Jude Lubega2 and Gilbert Maiga3
1
School of Computer Science and Information Technology, Kampala University,
Kampala, Uganda
[email protected]
2
3
Uganda Technology and Management University,
Kampala , Uganda
[email protected]
Department of Information Technology, College of Computing and Information Sciences,
Makerere University, Kampala , Uganda
gilmaiga@yaho o.com
technology revolution. Research shows that demand for
higher education is expanding [1] exponentially
worldwide, and about 150 million people are estimated to
seek for tertiary education by the year 2025.
Abstract
Adoption of multimedia based e-learning approach in higher
education institutions is becoming a key and effective component
for curriculum delivery. Additionally, multimedia teaching and
learning approach is a powerful tool to increase the perceived level
of user satisfaction, leading to enhance the blended learning
process. However, transmission of multimedia content over the
Internet in Constrained Bandwidth Environment (CBE) is a
challenge. This paper developed a framework that addressed the
most critical issue not addressed by the existing blended learning
frameworks after careful literature scanning. The framework was
tested and validated through experimental scientific research
process. This paper summarizes the main findings of the PhD
Thesis defended by the author some time back.
On the other hand, a larger percentage of eligible
candidates do not get access to higher education in LDCs.
According to Millennium Development Goal [2], the
university enrollment in Sub-Sahara Africa is the lowest
as compared to other parts of the world.
To be part of the global village, it is necessary for SubSaharan countries to make transition from natural
resources based economies to knowledge based
economies, in an effort to make departure from the
existing situation. Researchers argue that e-learning can
be used to deliver cost effective, flexible, and quality
higher education more effectively with improved access
by learners [3] and has the capacity to provide education
for all in LDCs [4]. However, universities in LDCs are
integrating Information Communication Technologies
(ICT) in their systems at a very slow pace. UNESCO
Institute for Statistics report [5] states that “in SubSaharan Africa, it is estimated that only 1 in 250 people
have access to the Internet as against the global average of
1 in 15” [6].
Keywords: Blended Learning, Framework, Multimedia
learning, Optimization, Constrained Bandwidth Environment.
.
1. Introduction
In the current knowledge driven society, the developments
in Information Technology have brought about many
positive changes in every field including education.
Driven by the growing demand of higher education
together with increasingly available new innovative
information technologies to increase access to learning
opportunities, the organizations in Lease Developed
Countries (LDCs) are considering cost effective and
efficient means for course delivery [34]. The aim of large
scale moves to use technology to deliver educational
content, by organizations in emerging environments is to
prepare them for the information and communication
The implementation of e-learning is an ideal choice for
the universities. And the introduction of the concept of
blended learning model is a natural start to achieve the
goal for the universities in developing countries, which is
pursued by many organizations in the developed world.
130
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
2. What type of multimedia content is compatible with the
challenges of Constrained Bandwidth Environment?
3. What are the factors of e-readiness leading to the
development of an e readiness model for the organizations
seeking to adopt blended learning systems, and their
challenges within Constrained Bandwidth Environment?
4. What framework design supports the implementation of
blended learning process in Constrained Bandwidth
Environment?
5. How well does the proposed framework address the
issues of Constrained Bandwidth Environment?
Research indicates that blended learning approach is the
most effective teaching and learning model [7], [8].
Although the concept of blended learning can prove as a
ground breaking and innovative approach in the context
of developing countries [9], it is not a panacea. There are
some obstacles such as Constrained Bandwidth
Environment (CBE) which can be a key challenge to its
adoption in the sub-region [10]. Constrained Bandwidth
Environment refers to insufficient bandwidth as compared
to user needs, coupled with other constraints such as high
cost, misuse and mismanagement due to ineffective or
non-existent Bandwidth Management Policies (BMP),
viruses and spam etc.
3. Methodology
This study used mixed research methods; qualitative and
quantitative that had approaches such as case study,
surveys, and experiments, dictated by the nature of the
study and research questions that could not be answered
by using a single approach [15].
A large number of organizations in the world are
increasingly adopting blended learning solutions in their
systems and several frameworks have been designed by
various authors such as [10], [11], [12], [13], [14]. A
critical review of the issues addressed by the existing
blended learning frameworks suggests that many aspects
of constrained bandwidth environment are taken for
granted that need to be addressed explicitly.
3.1. Research Design Framework
This research followed the generalized research design
framework format as suggested by [15]. Fig. 1 presents
the research design framework of the study.
2. Aim of the Research
The main objective of this research was to develop a
framework for adoption of blended learning process in
Constrained Bandwidth Environment, a peculiar state
associated with most of the developing countries.
Specifically, the study intended:
(a) To conduct a critical investigation of the literature on
current approaches aimed to identify gaps and challenges
for implementing blended learning process, within
Constrained Bandwidth Environment.
(b)To identify multimedia (learning content) compatible
with the challenges of Constrained Bandwidth
Environment.
(c) Exploring factors of e-readiness leading to the
development of an e-readiness model for organizations
seeking to implement blended learning process and their
challenges in the context of Constrained Bandwidth
Environment.
(d) To design a framework for implementing blended
learning process in Constrained Bandwidth Environment.
(e) To test and validate the designed framework.
Fig. 1 Research design framework
To address the specific objectives, this research sought to
develop the framework guided by the following research
questions:
1. What are the current approaches and their challenges
for implementing blended learning process, within
Constrained Bandwidth Environment?
First phase presented the ‘Critical Literature Review’ to
identify Research Question.
Second phase was
‘Gathering and Analyzing user Requirement Data’ for elearning which is part of blended learning process. Third
phase was ‘Design of the Framework’. Design of the
131
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
and empirical investigation. Literature investigation
aimed to identifying multimedia compatible with
challenges of CBE and focused on the most important and
dominant multimedia application, namely video format,
using a Multi-Level Systematic Approach (MLSA). The
research found that MPEG4 (.mp4) is the only video
format, in addition to MP3, JPEG, the audio and graphic
formats that are compatible with the challenges of
developing countries, to answer Research Question 2.
framework was informed by first and second phases that
worked as input to the design phase. The designed
framework was tested and validated at fourth phase that
completed the framework development process. Following
sections present the brief description of all the phases.
3.1.1 Identifying approaches for adoption of Blended
learning process in CBE
The aim of identifying existing blended learning
approaches, strategies, and their challenges was to
identify the research question. To achieve the objective,
literature survey was conducted by using various ‘key
words’ and ‘citations’ during research process aimed to
provide an answer to Research Question 1. Secondary
source of data was acquired through the studying of
technical literature; research articles, research reports,
books, journals, reports on the general problem area,
attending and participating in International conferences,
interaction with e-learning experts, and stakeholders [16].
The literature survey also included relevant PhD thesis
reports submitted to various universities in the world.
Aligning a strategy with challenges, opportunities, and
realities that abound the context is an important
preliminary step leading to the successful adoption of an
innovation [19]. Also the process of development of such
a strategy must involve the concern of relevant
stakeholders at the earlier stage on one hand and the
stakeholders must be ‘e-ready’ on the other hand, in order
to design a coherent achievable strategy tailored to meet
their needs [20]. The key stakeholders identified by this
research included: Institutions, lecturers, and students
[21].
Basing on this background, this research used a multiple
case study approach, an appropriate research method in
the situation when the intent of the study is description or
theory building as suggested in [22], in order to achieve
the underlying objective of the empirical investigation.
Researchers selected 3 universities in Uganda, in line with
[23]
guidelines
and
developed
three
survey
questionnaires; Institutional Readiness Questionnaire
(IRQ), Lecturers’ Perception Questionnaire (LPQ), and
Students’ Perception Questionnaire(SPQ) to explore
factors contributing to e-readiness in constrained
bandwidth environment as shown in Fig.2.
The literature review results revealed that [10], [11], [12],
[13], [14] are most popular models for adoption of blended
learning. An overview of the identified blended learning
frameworks suggest that mainly they address issues related
to the following components: Institutional (Administrative
& academic affairs etc.); Learners & Lecturers concerns
(Audience requirement analysis); Pedagogical (Content
analysis). Technological Infrastructure (Hardware &
software); Learning Content (Teaching and learning
material); Management & Evaluation (Logistics, student
support services); Ethical (Copyright etc.); Resources
(Human & financial resources); Interface Design (user
interface); and Interactivity (Interactivity among various
learning modes in blended learning environment).
Prior reliability testing of the data collection tools was
done using Chronbach’s alpha coefficient, in line with
suggestions of [24] that gave the average correlation
among the various items in the questionnaires. The
responses were coded into the Statistical Package for
Social Sciences (SPSS) to run the reliability test. The
Chronbach’s alpha coefficient value for Institutional and
Students’ questionnaires were found above 0.7 which is
acceptable. However, for lecturer’s questionnaire,
reliability test could not run because most of the questions
were open ended aimed to extract information from
lecturers rather than seeking their opinion.
The critical analysis of literature points to the lack of
frameworks for adaptation of blended learning process
that addresses the important issue of constrained
bandwidth environments. The need has therefore
remained for framework that support implementation of
blended learning process in constrained bandwidth
environments [18].
The reviewed literature gave a way forward to gather data
on user requirements.
3.1.2 Gathering User Requirements Data
User requirements data was gathered through study
investigation that comprised of literature investigation
132
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
bandwidth environment. E-learning which is an essential
part of blended learning occurs, when at least there is
learning content (multimedia) and network to deliver the
content [27]. Hence multimedia and network (bandwidth)
related data constitutes the main part of user requirements
data for design of blended framework adaptable in
constrained low bandwidth environment.
On that pretext, section 3.1.2 gathered data on user
requirements; identified multimedia (learning content)
compatible with challenges of constrained bandwidth
environment through literature investigation and identified
the most critical (bandwidth or network
related)
challenges, organizations seeking to implement blended
learning solutions are facing. The framework that addressed
the issue of constrained bandwidth environment not
addressed by existing blended learning frameworks is
presented in Fig.3. The framework presented used a
popular framework by Khan [13] as basis and has two main
components; Network Optimization and Multimedia
Optimization.
Fig.2 Survey research design
The questionnaires were distributed among identified
stakeholders in a sample of three organizations
(universities); A, B and C. The study applied convenient
sampling technique used widely in Doctoral Thesis
reports (e.g., [25]). Furthermore the methodology used is
appropriate for this research and has achieved recognition
by researchers at International level (e.g., see [26]).
The study concluded with identification of the network
(bandwidth) related issues and assessed the level of ereadiness leading to the development of an e-readiness
model, influenced by three survey questionnaires; IRQ,
LPQ, and SPQ as the answer to Research Question 3.
Lack of sufficient bandwidth, Low speed Internet, High
cost of bandwidth, Ineffective or non-existent Bandwidth
Management Policies (BMP) leading to mismanagement
of available little bandwidth were among the main critical
issues identified during the investigation.
Fig. 3 Optimization Based Blended Learning Framework for (CBE)
(OBBLFCBE)
Identification of multimedia compatible with challenges
of constrained bandwidth environment and an E-readiness
Model, was a step towards designing a framework that
addressed the critical issues not addressed by existing
blended learning frameworks and models.
Khan’s framework has eight dimensions; Institutional,
Pedagogical, Technological, Interface design, Evaluation,
Management, Resource support, and Ethical. Although
Khan’s framework considered many factors leading to the
adoption of a successful blended learning strategy in
organizations. However, the issue of adopting blended
learning process in constrained bandwidth environment
has not been addressed. More particularly, the
Pedagogical dimension of the framework states that
organizations are required to provide sufficient
3.1.3 Design of the Framework
Design of the framework is informed by previous sections
3.1.1 and 3.1.2. Section 3.1.1 affirmed the need of a
blended learning framework adaptable in constrained
133
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
bandwidth, which is limited in CBE, and carryout
learning content analysis without providing details about
the type of content compatible with the challenges of
constrained bandwidth environment where designing the
content is costly.
The underlying thesis defended in this research addressed
the above most two critical issues and posits that: in
constrained bandwidth environment, optimizing the
network and multimedia content supports a framework for
the adoption of blended learning process. Moreover,
research shows that application and network are symbiotic
and performance is enhanced when both network and
learning content are optimized together [28]. These two
issues were impeding the technology integration process
in developing countries.
The framework as shown in the Fig.3 comprises of a
sequence of processes. Each process has its input and
output. Network
optimization and Multimedia
optimization are the main components governed by their
respective Quality of Service (QoS) rules. Preoptimization states are inputs to optimization states of
network and multimedia processes, whereas postoptimization states act as their outputs.
Fig. 4 Network Optimization
Authentication: User Authentication System (UAS) can
be used to restrict the access to network resources from
unauthorized users by issuing them log in passwords.
[30].
However, for next process the interaction between the
optimized states of network and multimedia, these two
outputs act as inputs, and output results of this process
provide a high performing multimedia content to be used
to implement blended learning process in constrained
bandwidth environment and close the loop. The details of
the main components of the framework; Network
Optimization and Multimedia Optimization along with
various related technological concepts are discussed in the
following sections.
Prioritizing the network traffic: Prioritization technique
is applied to enhance the Quality of Service (QoS) for
real time applications such as video and audio by
processing them before all other data streams, aiming to
provide low latency as dictated by policy rules configured
on the router interface. In a similar manner, the
technique is used to prioritize among various categories of
users based upon their associated privileges [30].
3.1.3.1 Network Optimization
Time-Based ACLs: Time based ACLs are used to restrict
the users from downloading and uploading huge files,
during the busy working hours of the week that consumes
a lot of bandwidth. ACLs can be configured on a router in
the network system by using simple command (Cisco IOS
Software Release 12.0.1.T) by creating the time range on
the router, using preferably the Network Time Protocol
(NTP) synchronization.
Fig. 4 as mentioned in Fig. 2 outlines the techniques that
can be used to optimize the network system in an
organization. Pre-optimization and Post-optimization
states of network are also shown in the figure for the
purpose of comparison.
Filtering the network traffic that is done using packet
filter firewalls is aimed to reduce the congestion. The
acceptance or discarding decision of each arriving packet is
made by the firewall by comparing the information at its
header, with a set of governing rules called Access Control
Lists (ACLs) [29].
Cache and Proxy: Cache and proxy, the two local
memory devices are the best option to save bandwidth
[31]. The cache saves the most frequently digital contents
needed by the students. On the hand, proxy servers are
used to save the recently or more often visited WebPages.
134
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
provides better quality at low bit rate, is appropriate for use
in developing countries.
3.1.3.2 Multimedia Optimization
Fig.5 as mentioned in Fig.3 presents the multimedia
optimization techniques along with Pre-optimization and
Post-optimization states.
Audio: Research indicates that MPEG-3 (MP-3) audio
format is compatible with the challenges of CBE [34].
Graphics: Joint Photographic Expert Groups (JPEG) and
Portable Network Graphics (PNG) graphic files are
identified as compatible with CBE [34].
3.1.4 Framework Testing and Validation
3.1.4.1 Framework Testing
An experimental testbed as shown in figure 6 was
designed to test and validate the designed framework.
Fig.5 Multimedia Optimization
Multimedia may refer to presentation of information
using multiple forms of media (text, graphics, audio, and
video). The concept of interactive multimedia in
constrained bandwidth is still a challenge [20]. However,
compression and streaming techniques stated in Fig.5 can
be used to enhance the media quality [32].
Multimedia
Streaming:
Multimedia
streaming
techniques enable the user to receive a real time continuous
audio video data once the sender start sending the file,
without waiting when full contents are ready in local
storage. However, the data is compressed before it is
transmitted and is again decompressed before viewing by
the viewer. In on demand option, the client receives the
streaming multimedia contents as per demand.
Fig.6 Experimental test bed system architecture
The testbed used following main hardware and software;
two computers connected to a Cisco router by means of a
Cisco switch, 128 kb broad band Internet connection,
various multimedia files (video, audio, graphics, and
text), multimedia conversion software, and “Net catcher”a multimedia performance measuring tool installed on
one of the client computers. We acquired one video file
from Internet and converted it to various other types for
the sake of consistency.
Multimedia Compression: The compression technique is
applied to reduce the huge file size to manageable level for
efficient transmission of multimedia content. The
compression process significantly reduces the bandwidth
requirements as required in the context of LDCs.
Afterwards, all these files were compressed-converted to
.mp4 (commonly used video standard). Similar procedure
was adopted for other types of media files. The userdriven metric ‘latency’ was used to measure the
performance of various media (video, audio, graphics, and
text) files when they were transmitted over a wireless
network connection. The media performance was
measured in four phases:
Video. Transmission of time sensitive multimedia contents
such as video requires much bandwidth is challenge in low
bandwidth networks. However, the compression technique
applied under Quality of Service (QoS) can significantly
reduce the demand for high bandwidth [33]. Section 3.1.2
identified MPEG4 (MP4) as the standard video format
compatible with constrained bandwidth environment that
135
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Phase I:
Unoptimized Network & Unoptimized
Multimedia (UNUM)
Phase II: Unptimized Network & Optimized Multimedia
(UNOM)
Phase III: Optimized Network & Unoptimized
Multimedia (ONUM)
Phase IV: Optimized Network & Optimized Multimedia
(ONOM)
Fig. 9 Latency results for unoptimized flv
Multimedia Performance for Unoptimized Network:
This section presents the screen shots taken by Net
Catchlite tool that measured the latency, when
unoptimized and optimized multimedia (.flv video) files
were transmitted over the unoptimized wireless network.
The average latency for unoptimized and optimized .flv
file was 559ms and 422ms as shown in Fig.7and Fig.8
respectively.
Fig. 10 Latency results for optimized .flv
Analysis of Multimedia performance results: The
analysis of .flv video format performance is presented in
Table 1 that illustrates the latency comparison for
various phases.
Table 1 Latency results analysis for .flv
Fig.7 latency results for unoptimized .flv video file
The percentage of latency reduction is also calculated. We
notice that at every phase, latency factor continue to
reduce as compared to initial phase when both network
and multimedia files were not optimized. The results
further indicate that at phase IV when both network and
multimedia were optimized, latency factor reduced
significantly, from 559 ms to 212, suggesting a reduction
of 62.07%.
Fig. 8 latency results for optimized .flv video file
Multimedia Performance for Optimized Network: Prior
to transmission of multimedia files over the network, this
research optimized network by configuring policy map on
the router interface by allocating specific amounts of
bandwidth to various classes and prioritizing a specific
type (e.g., video and audio) over other network traffic,
aimed to apply Quality of Service (QoS) to increase the
perceived level of user satisfaction [35].
In a similar manner, 6 other video files (.mov, .mkv,
.mpg, .mpg2, .avi, .mp4); 3 Audio files (WAV, M4A,
MP3); 3 Graphic file files (bmp, tiff, JPEG); and 2 Text
(PPT) files were transmitted over Internet one by one
and results calculated in 4 phases. Study results indicated
that performance of multimedia was improved
significantly, when network and application were both
optimized that verified the main theoretical concept of the
framework.
With optimized network state, Net Catchlite tool
measured the latency of unoptimized and optimized
multimedia (video) files (.flv) when transmitted over the
optimized wireless network. The average latency for .flv
was 221ms and 212 ms respectively as shown in Fig 9 and
10.
136
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
3.1.4.2 Framework Validation
The novelty of this research is that it linked practical with
theoretical foundation and relevant literature, aimed to
make both scientific and practical contribution, and laid
the foundation stone for further research in the area to
make subsequent developments[17].
To ensure the soundness and accuracy of the designed
framework, Verification and Validation (V&V) was done
based on: Correspondence between model and statement
of the problem which is in line with results presented in
above section; Model Behavior was tested and proved that
it does what it is expected to do; and Conservation of
Flow was confirmed from consistent results in section
3.1.4 when several media files from each category were
transmitted over the network that shows the stability of
the model. The validation process was further enhanced
by sending prototype results to users and IT staff working
in various universities in Uganda. The designed
framework was highly evaluated by users and IT
professionals as shown in Table 2 (1- lowest score, 4highest score) as answer to Research Question 5.
The output of this study has several far reaching
implications for community of stakeholders that include;
higher education institutions, students, researchers, and
information system providers.
This research addresses the question of high practical
value, is rooted into realistic and idealistic grounds that
leads to increased usefulness of knowledge created to
guide researchers. The study focused on evaluating the
performance of learning content which is an integral part
of blended learning process.
Table 2: Framework evaluation results by stakeholders
This research can make significant improvements on the
landscape of African higher education system. At the
same time, this study has given a way forward to
Institutions of Higher learning to protect their valuable
institutional resource (bandwidth) through prioritizing
time-sensitive applications to minimize delays. The
findings of the study can be used by information system
developers,
e-learning
solution
providers,
and
instructional designers to increase the perceived level of
user–satisfaction when delivering multimedia content to
the users in LDCs.
The future research may include:
(i)How to improve the performance of multimedia
content, to increase further the level of user satisfaction
by reducing latency factor more than shown in the
current study?
(ii)How the findings of this research can be replicated in
different context?
(iii)The research utilized mixed methods approach that
enabled the researcher to study depth and breadth of the
problem, provided the ability to triangulate the findings of
the research methods used and hence giving a way
forward for the utilization of the methodology by other
researchers in future studies.
(iv)Researchers can use Multi Level Systematic Approach
(MLSA) introduced in this study for in-depth analysis of
research problems.
4. Conclusions
The fundamental contribution of this research was
providing Optimization Based Blended Learning
Framework for Constrained Bandwidth Environment,
with multimedia optimization and network optimization
as its two main components. The research achieved this
main objective. The framework emphasizes that
performance of multimedia (learning content) can be
improved when network and application are both
optimized, in return that would increase the perceived
level of user satisfaction, enhancing the blended learning
process for organizations seeking to implement blended
learning solutions in developing countries. Increasing the
perceived level of user satisfaction in developing
economies is a big mile stone indeed [17].
References
[1] B. A. Digolo, E.A. Andang J. Katuli, “E- Learning as a
Strategy for Enhancing Access to Music Education”, International
Journal of Business and Social Science Vol. 2 No. 11 [Special
Issue - June 2011].
[2]Millinuim Development Goal Report, 2010.
137
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Heidelberg, ISBN: 978-3-642-22762-2, Vol. 6837, 2011, pp.
188-199.
[18]L. Casanovas, “Exploring the Current Theoretical
Background About Adoption Until Institutionalization of Online
Education in Universities: Needs for Further Research”, Electronic
Journal of e-Learning, Vol. 8, Issue 2, 2010, pp. 73-84.
[19]CANARIE report, “Framework for Rural and Remote
Readiness in Telehealth” June, 2002.
[20] A. Sife, E. Lwoga, and C. Sanga, "New technologies for
teaching and learning: Challenges for higher learning institutions
in developing countries", International Journal of Education and
Development using ICT, Vol. 3, No. 2, 2007, pp. 57-67.
[21]M. Faridha, “A Framework for Adaptation of Online
Learning. In Towards E-nhancing Learning With Information and
Communication Technology in Universities”, A Master’s thesis in
computer science Makerere University, 2005.
[22] R. K. Yin, “Case study research: Design and method”,
Newbury Park, CA: Sage Publication, 1984.
[23]J. Creswell, “Educational research: Planning, conducting, and
evaluating quantitative and qualitative research”, Upper Saddle
River, NJ: Merrill Prentice Hall, 2002.
[24] D. M. Gabrielle, “The effects of technology-mediated
instructional strategies on motivation, performance, and self
directed learning”, PhD Thesis Florida State University College
of Education, 2003.
[25] M.C. Wesley, “Family and Child Characteristics Associated
with Coping Psychological Adjustment and Metabolic Control in
Children and Adolescents with Type 1 Diabetes” , A PhD Thesis
report The University of Guelph, April, 2012.
[26] N. A. Suhail, “Assessing Implementation of Blended learning
in Constrained Low Bandwidth Environment” In Aisbet, J.,
Gibbon,G., Anthony, J. R., Migga, J. K., Nath, R., Renardel, G.
R. R.: Special Topics in Computing and ICT Research.
Strengthening the Role of ICT in Development, August, 2008,
Vol. IV, pp. 377-390, Fountain publishers Kampala,Uganda
[27]C. Snae and M. Bruckner, “Ontology driven e-learning
system based on roles and activities for Thi learning
environment”, Interdisciplinary Journal of knowledge and
Learning Objects, Vol. 3, 2007.
[28]Understanding Network-Transparent Application Acceleration
and WAN Optimization. Cisco white paper, 2007.Cisco systems.
Inc.
[29] U. Thakar, L. Purohit, and A. Pahade, “An approach to
improve performance of a packet filtering firewall”, Proceedings
of wireless and optical communications networks, 2012 Ninth
International Conference on 20-22 September, 2012, Indore, pp.15.
[30] M. Kaminsky, “User Authentication and Remote Execution
Across Administrative Domains”, Doctor of Philosophy Thesis,
Massachusetts Institute of Technology, 2004.
[31]Z. Xu, L. Bhuyan , and Y. Hu, “Tulip: A New Hash Based
Cooperative Web Caching Architecture”, The Journal of
Supercomputing, 35, 2006, pp. 301–320.
[32]N.A. Suhail and E.K. Mugisa, “Implementation of E-learning
in Higher Education Institutions in Low Bandwidth Environment:
A Blended Learning Approach”, in Kizza J.M., Muhirwe J.,
Aisbett J., Getao K., Mbarika V., Patel D., and Rodrigues A. J.:
Special Topics in Computing and ICT Research Strengthening the
URL:http://www.un.org/millenniumgoals/pdf/MDG%20Report%
202010%20En%20r15%20-low%20res%2020100615%20-.
[3]A.W. Bates, “Technology, Open Learning and Distance
Education” Routledge London, 1995.
[4] S. O. Gunga, I. W. Ricketts, “Facing the challenges of elearning initiatives in African Universities”, British Journal of
Educational Technology Vol. 38, No. 5, 2007.
[5] UNESCO Institute for Statistics, Global education digest 2006:
Comparing education statistics across the world.
URL: http://www.uis.unesco.org/ev.php?ID=6827 201&
ID2=DOTOPIC
[6] S. Asunka, H.S. Chae, “Strategies for Teaching Online
Courses within the Sub-Saharan African Context: An
Instructor's Recommendations”, MERLOT Journal of Online
Learning and Teaching Vol. 5, No. 2, June 2009.
[7] F. Alonso, G. Lopez, D. Manrique, and J. M. Vines, “An
instructional model for web based education with blended learning
process approach” , British Journal of Educational Technology,
36(2), 2005, pp. 217-235.
[8]E. Nel, “Creating meaningful blended learning experience in a
South Africa classroom: An Action inquiry”, Thesis submitted in
fulfillment of the requirements for the degree Philosophiae Doctor
in Higher Education Studies, University of The Free State, 2005.
[9]K. P. Dzvimbo, “Access Limitation and the AVU Open,
Distance and e-Learning (ODeL) solutions”, Proceedings of The
Association of African Universities conference Cape Town,
February 22, 2005.
[10] Bersin and associates, “4-factor Blended Learning
Framework, What Works in Blended Learning”, ASTDs Source of
E-learning, Learning Circuits 2003. http://www.bersin.20.com/
510.654.8500
[11] J. De Vries, “E-learning strategy: An e-learning Framework
for Success”, Blue Streak learning, Quality learning solutions,
Bottom line results. 2005.
elearningcentre.typepad.com/whatsnew/2005/08/elearning
strat.html
[12] A Blended Learning Model. In Training: A Critical Success
Factor in Implementing a Technology Solution”, North America
Radiant Sytems, 2003, Inc. 3925 Brookside Parkway Atlanta, GA
30022. URL: www.radiantsystems.com (Accessed on 28 June,
2007)
[13] B. H. Khan, “A framework for Web-based authoring
systems”, Web-based training, 2001, pp.355-364, Englewood
Clffs, NJ: Educational technology Publications.
http://iteresources.worldbank.org/EXTAFRREGTOPDISEDU/Re
sources/Teacher education Toolkit
[14], F. Z. Azizan, “Blended Learning in Higher Institutions in
Malaysia”, Proceedings of Regional Conference on Knowledge
Integration in ICT, 2010.
[15] K. Mc Dougall, “A Local-State Government Spatial Data
Sharing Partnershi Model to Facilitate SDI Development”, PhD
Thesis, University of Melbourne, 2006.
[16] A. Strauss, J. Corbin, “Basics of qualitative research:
Grounded theory procedures and techniques”, London: Sage.
1990.
[17] N. A. Suhail, J Lubega, and G Maiga, “Optimization Based
Multimedia Performance to Enhance Blended Learning
Experience in Constrained Low Bandwidth Environment”,
Lecture Notes in Computer Science, Springer-Verlag Berlin
138
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Role of ICT in Development, Vol. III, August, 2007, pp. 302-322,
Fountain publishers, Kampala, Uganda.
[33] S. Mishra and U.Sawarkar, “Video Compression Using
MPEG”, Proceedings of International Conference & Workshop on
Recent Trends in Technology, proceedings published in
International Journal of Computer Applications (IJCA), 2012
[34]N. A. Suhail, J. Lubega,, and G. Maiga, “ Multimedia to
Enhance Blended Learning Experience in Constrained Low
Bandwidth Environment”, Lecture Notes in Computer Science,
Springer-Verlag Berlin Heidelberg, ISBN: 978-3-642-320170,Vol.7411, 2012, pp. 362—372.
[35]A. A. Abbasi, J. Iqbal, and A. N. A. Malik, “Novel QoS
Optimization Architecture for 4G Networks”, in World Academy
of Science, Engineering and Technology, 2010.
139
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Analysis of Computing Open Source Systems
J.C. Silva1 and J.L. Silva2
1
Dep. Tecnologias, Instituto Politécnico do Cávado e do Ave,
Barcelos, Portugal
[email protected]
2
Madeira-ITI, Universidade da Madeira
Funchal, Portugal
[email protected]
implementation. Each widget has a fixed set of properties
and at any time during the execution of the GUI, these
properties have discrete values, the set of which
constitutes the state of the GUI. The programmers interest,
besides satisfying the user, is in the intrinsic quality of the
implementation, which will impact the system's
maintainability.
As user interfaces grow in size and complexity, they
become a tangle of object and listener methods, usually all
having access to a common global state. Considering that
the user interface layer of interactive open source systems
is typically the one most prone to suffer changes, due to
changed requirements and added features, maintaining the
user interface code can become a complex and error prone
task. Integrated development environments (IDEs), while
helpful in that they enable the graphical definition of the
interface, are limited when it comes to the definition of the
behavior of the interface.
In this paper we explore an approach for the analysis of
open source system's user interfaces. Open-source
software is software whose source code is made available,
enabling anyone to copy, modify and redistribute the
source code without paying royalties or fees. This paper
discusses an approach to understand and evaluate an open
source system from an interactive perspective. We present
a static analysis based framework for GUI-based
applications analysis from source code.
In previous papers [1,3] we have explored the
applicability of slicing techniques [4] to our reverse
engineering needs, and developed the building blocks for
the approach. In this paper we explore the integration of
analysis techniques into the approach, in order to reason
about GUI models.
The paper is organized as follow: Section three discusses
the value of inspecting source code from a GUI quality
perspective; Section four introduces our framework for
GUI reverse engineering from source code; sections five
Abstract
Graphical user interfaces (GUIs) are critical components of
today's open source software. Given their increased relevance,
the correctness and usability of GUIs are becoming essential.
This paper describes the latest results in the development of our
tool to reverse engineer the GUI layer of interactive computing
open source systems. We use static analysis techniques to
generate models of the user interface behavior from source code.
Models help in graphical user interface inspection by allowing
designers to concentrate on its more important aspects. One
particular type of model that the tool is able to generate is state
machines. The paper shows how graph theory can be useful
when applied to these models. A number of metrics and
algorithms are used in the analysis of aspects of the user
interface's quality. The ultimate goal of the tool is to enable
analysis of interactive system through GUIs source code
inspection.
Keywords: analysis, source code, quality.
1. Introduction
In the user interface of an open source software systems,
two interrelated sets of concerns converge. Users interact
with the system by performing actions on the graphical
user interface (GUI) widgets. These, in turn, generate
events at the software level, which are handled by
appropriate listener methods. In brief, and from a user's
perspective, graphical user interfaces accept as input a
pre-defined set of user-generated events, and produce
graphical output. The users' interest is in how well the
system supports their needs.
From the programmers perspective, typical WIMP-style
(Windows, Icon, Mouse, and Pointer) user interfaces
consist of a hierarchy of graphical widgets (buttons,
menus, text-fields, etc) creating a front-end to the software
system.
An event-based programming model is used to link the
graphical objects to the rest of the system's
140
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
conditions in high order logic. Proof of the conditions is
usually conducted within interactive theorem provers such
as PVS or Coq [11,12].
We believe that defining and integrating a methodology
into open source systems development processes should
be the first priority to certificate open source systems.
and six presents the analysis of a software system; Section
seven discusses the results of the process; the paper end
with conclusions in Section eight.
2. Analysis of Open Source Systems
Open source systems are popular both in business and
academic communities with products such as Linux,
MySQL, OpenOffice or Mozilla. Open source systems are
free redistribution with source code accessible and
complying several criterions. The program must allow
distribution in source code as well as compiled form.
Deliberately obfuscated source code is not allowed.
Intermediate forms such as the output of a preprocessor or
translator are not allowed. The license must allow
modifications and derived works, and must allow them to
be distributed under the same terms as the license of the
original software [5]. Considering that open source
systems are typically prone to suffer changes, due to
modifications and derived works, maintaining the system
and its usability can become an error prone task [6]. A
number of challenges remain to be met, however, many of
which are common to all open source projects. This
Section discusses open source systems analysis as a way
to foster adoption and deployment of open source systems.
The objective of the open source analysis is to evaluate
the quality of open source systems involving software
analysis and engineering methodologies. In the literature,
several directions are used for achieving this goal such as
testing, light weight verification and heavy weight
verification, e.g [7,8]. Testing is a huge area for open
source analysis [9]. Different kinds of tests are applied
such as functional testing, regression testing, stress testing,
load testing, unit testing, integration testing,
documentation analysis, source code analysis, reverse
engineering. Lightweight verification includes various
methods of static analysis and model checking, e.g. [10].
These may include identification of domain specific
restrictions and typical bugs for automatic detection,
formal representation of the restrictions in terms of the
tools used, development of simplified models of target
system to be used for automatic analysis, automatic
analysis of target source code with verification tools and
investigation and classification of results.
Another approach is heavyweight verification providing a
more complete analysis of the quality of the source code
system. There are different approaches to heavyweight
verification. Classical methods of verification requires to
formally describe requirements in the form of
precondition and post-condition. Then, invariants and
variants should be defined for the open source system.
After that verification tools automatically generate
3. Inspection from source code
The evaluation of an open source software is a
multifaceted problem. Besides the intrinsic quality of the
implementation, we have to consider the user reaction to
the interface (i.e. its usability [13]). This involves issues
such as satisfaction, learnability, and efficiency. The first
item describes the user's satisfaction with the open source
system. Learnability refers to the effort users make to
learn how to use the application. Efficiency refers to how
efficient the user can be when performing a task using the
application.
The analysis of a system's current implementation can
provide a means to guide development and to certify
software. For that purpose adequate metrics must be
specified and calculated [14,15]. Metrics can be divided
into two groups: internal and external [16]. External
metrics are defined in relation to running software. In
what concerns GUIs, external metrics can be used as
usability indicators. They are often associated with the
following attributes [17]:
•
•
•
•
•
Easy to learn: The user can do desired tasks
easily without previous knowledge;
Efficient to use: The user reaches a high
productivity level.
Easy to remember: The re-utilization of the
system is possible without a high level of effort.
Few errors: Errors are made hardly by the users
and the system permits to recover from them.
Pleasant to use: The users are satisfied with the
use of the system.
However, the values for these metrics are not typically
obtainable from direct analysis of the implementation,
rather through users' feedback to using the system.
Internal metrics are obtained by source code analysis, and
provide information to improve software development.
Such metrics measure software aspects, such as source
lines of code, functions invocations, etc. A number of
authors has looked at the relation between internal metrics
and GUI quality. Stamelos et al. [18] used the Logiscope 1
tool to calculate values of selected metrics in order to
1http://www-01.ibm.com/software/awdtools/logiscope/
141
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
supported by a given GUI implementation.
The tool performs the parsing of the source code. A
module executes this step. To implement this first module,
a parser for the programming language being considered
is used. The tool has been used to reverse engineer Java
and Haskell [21] programs written using the (Java) Swing,
GWT, and (Haskell) WxHaskell GUI toolkits. For the
Java/Swing and GWT toolkits, the SGLR parser has been
applied whose implementation can be accessible via the
Strafunski tool [22]. For the WxHaskell toolkit the
Haskell parser that is included on the Haskell standard
libraries was used. Whatever the parser, it generates an
Abstract Syntax Tree (AST). The AST is a formal
representation of the abstract syntactical structure of the
source code.
The full AST represents the entire code of the application.
However, the tool's objective is to process the GUI layer
of interactive open source systems, not the entire source
code.
To this end, an another module implements a GUI code
slicing process using strategic programming. The module
is used to slice the AST produced by the compiler, in
order to extract its graphical user interface layer. The
module is composed of a slicing library, containing a
generic set of traversal functions that traverse any AST.
Once the AST has been created and the GUI layer has
been extracted, GUI behavioral modeling can be
processed. A module implements a GUI abstraction step.
The module is language independent. It generates a model
of user interface behavior. The relevant abstractions used
in the model are user inputs, user selections, user actions
and output to user.
More specifically, the modules generates GUI-related
metadata files with information on possible GUI events,
associated conditions and actions, and states resulting
from these events. Each of these items of data are related
to a particular fragment from the AST. These are GUI
specifications written in the Haskell programming
language. These specifications define the GUI layer by
mapping pairs of event/condition to actions.
study the quality of Open Source code. Ten different
metrics were used. The results enable evaluation of each
function against four basic criteria: testability, simplicity,
readability and self-descriptiveness. While the GUI layer
was not specifically targeted in the analysis, the results
indicated a negative correlation between component size
and user satisfaction with the software.
Yoon and Yoon [19] developed quantitative metrics to
support decision making during the GUI design process.
Their goal was to quantify the usability attributes of
interaction design. Three internal metrics were proposed
and defined as numerical values: complexity, inefficiency
and incongruity. The authors expect that these metrics can
be used to reduce the development cost of user interaction.
While the above approaches focus on calculating metrics
over the code, Thimbleby and Gow [20] calculate them
over a model capturing the behavior of the application.
Using graph theory they analyze metrics related to the
users' ability to use the interface (e.g., strong
connectedness ensure no part of the interface ever
becomes unreachable), the cost of erroneous actions (e.g.,
calculating the cost of undoing an action), or the
knowledge needed to use the system (e.g., the minimum
cut identifies the set of actions that the user must know in
order to to be locked out of parts of the interface).
In a sense, by calculating the metrics over a model
capturing GUI relevant information instead of over the
code, the knowledge gained becomes closer to the type of
knowledge obtained from external metrics. While
Thimbleby and Gow manually develop their models from
inspections of the running software/devices, an analogous
approach can be carried out analyzing the models
generated directly from source code. We have been
developing a tool to reverse engineer models of a user
interface from its source code [1,3]. By coupling the type
of analysis in [20] with our approach, we are able to
obtain the knowledge directly from source code. By
calculating metrics over the behavioral models, we aim to
acquire relevant knowledge about the dialogue induced by
the interface, and, as a consequence, about how users
might react to it. In this paper we describe several kinds of
inspections making use of metrics.
5. HMS Case Study: A Larger Interactive
System
4. The Tool
In previous Section, we have presented the implemented
tool. In this Section, we present the application of the tool
to a complex/large real interactive system: a Healthcare
Management System (HMS) available from Planet-sourcecode2, one of the largest public source code database on
the Internet. The goal of this Section is twofold: Firstly, it
is a proof of concept for the tool. Secondly, we wish to
The tool's goal is to be able to extract a range of models
from source code. In the present context we focus on finite
state models that represent GUI behavior. That is, when
can a particular GUI event occur, which are the related
conditions, which system actions are executed, or which
GUI state is generated next. We choose this type of model
in order to be able to reason about and test the dialogue
2 http://www.planet-source-code.com/
142
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
list [4]) which exit the system. These events can be
executed by clicking the Exit or Login buttons,
respectively. The informal description of login window
behavior provided at the start of the Section did not
included the possibility of exiting the system by pressing
the Login button. The extracted behavioral graph however
defines that possibility, which can occur if condition
cond2 is verified (cf. pair loginBtn/cond2 with action list
[4]). Analysing condition cond2 (source.equals(exitBtn)),
dead code was encountered. The source code executed
when pressing the Login button uses a condition to test
whether the clicked button is the Login button or not. This
is
done
through
the
boolean
expression
source.equals(loginBtn). However, the above action
source code is only performed when pressing the Login
button. Thus, the condition will always be verified and the
following else component of the conditional statement will
never be executed.
Summarizing the results obtained for the login window,
one can say that the generated behavioral graph contains
an event/condition/actions triplet that does not much the
informal description of the system. Furthermore, this
triplet cannot be executed despite being defined on the
behavioral model. This example demonstrates how
comparing expected application behavior against the
models generated by the tool can help understand (and
detect problems in) the applications' source code.
analyze the interactive parts of a real application.
The HMS system is implemented in Java/Swing and
supports patients, doctors and bills management. The
implementation contains 66 classes, 29 windows forms
(message box included) and 3588 lines of code.
The login window is the first window that appears to HMS
users. This window gives authorized users access to the
system and the HMS main form, through the introduction
of a user name and password pair. This window is
composed of two text box (i.e. username and password
input) and two buttons (i.e. Login and Exit buttons).
If the user introduces a valid user name/password and
presses the Login button, then the window closes and the
main window of the application is displayed. On the
contrary, if the user introduces invalid data, then a
warning message is produced and the login window
continues to be displayed. By pressing the Exit button, the
user exits the application.
Applying the tool to the source code of the application,
and focusing on the login window, enables the generation
of several models. Figure 1, for example, shows the graph
generated to capture the login window's behavior.
Associated to each edge there is a triplet representing the
event that triggers the transition, a guard on that event
(here represented by a label identifying the condition
being used), and a list of interactive actions executed
when the event is selected (each action is represented by a
unique identifier which is related to the respective source
code).
6. GUI Inspection through Graph Theory
This Section describes some examples of analysis
performed on the application's behavioral graph from the
previous section. We make use of the implemented tool
for the manipulation and statistical analysis of the graph.
6.1 Graph-tool
Graph-tool3 is an efficient python module for the
manipulation and statistical analysis of graphs. It allows
for the easy creation and manipulation of both directed or
undirected graphs. Arbitrary information can be associated
to the nodes, edges or even the graph itself, by means of
property maps. Graph-tool implements all sorts of
algorithms, statistics and metrics over graphs, such as
degree/property histogram, combined degree/property
histogram, vertex-vertex correlations, average vertexvertex shortest distance, isomorphism, minimum spanning
tree, connected components, maximum flow, clustering
coefficients, motif statistics, communities, or centrality
measures.
Fig. 1 HMS: Login behavioral graph
Analyzing this model, one can infer that there is an
event/condition pair (edge loginBtn / cond1, with action
list [1,2,3]) which closes the window (cf. edge moving to
close node). Investigating action reference 2, it can be
further concluded that another window (startApp) is
subsequently opened. Furthermore, one can also infer that
there are two event/condition pairs (edge exitBtn / cond4
with action list [6], and edge loginBtn / cond2 with action
3 http://projects.forked.de/graph-tool/
143
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
maximal distance from other points in the graph. Finding
the center of a graph is useful in GUI applications where
the goal is to minimize the steps to execute a particular
task (i.e. edges between two points). For example, placing
the main window of an interactive system at a central
point reduces the number of steps a user has to execute to
accomplish tasks.
Fig. 2 HSM: The overall behavior
Now we will consider the graph described in Figure 2
where all vertices and edges are labeled with unique
identifiers. Figure 2 provides the overall behavior of the
HMS system. This model can be seen in more detail in the
electronic version of this paper. Basically, this model
aggregates the state machines of all HMS forms. The right
top corner node specifies the HMS entry point, i.e. the
mainAppstate0 creation state from the login's state
machine (cf. Figure 1).
Fig. 3 HSM's pagerank results
6.2 GUI Metrics
PageRank is a link analysis algorithm, used by the Google
Internet search engine that assigns a numerical weighting
to each element of a hyperlinked set of documents. The
main objective is to measure their relative importance.
The wight assigned to each element represents the
probability that a person randomly clicking on links will
arrive at any particular page [23]. A probability is
expressed as a numeric value between 0 and 1. This same
algorithm can be applied to our GUI's behavioral graphs.
Figure 3 provides the result obtained when applying the
pagerank algorithm to graph of Figure 2. The size of a
vertex corresponds to its importance within the overall
application behavior. This metric can have several
applications, for example, to analyze whether complexity
is well distributed along the application behavior. In this
case, there are no particularly salient vertices, which is an
indication that interaction complexity is well distributed
considering the overall application. It is also worth
noticing that according to this criteria, the Main window is
As discussed in this paper, one of our goals is to show
how the implemented tool supports the use of metrics such
as those used by Thimbleby and Gow [20] to reason about
the quality of a user interface. To illustrate the analysis,
we will consider three metrics: Shortest distance between
vertices, Pagerank and Betweeness.
The Graph-Tool enables us to calculate the shortest path
between two vertices. This is useful to calculate the
number of steps to execute a particular task. These results
can be used to analyze the complexity of an interactive
application's user interface. Higher numbers of steps
represent complex tasks while lower values are
applications with simple tasks. It can also be applied to
calculate the center of a graph. The center of a graph is the
set of all vertices A where the greatest distance to other
vertices B is minimal. The vertices in the center are called
central points. Thus vertices in the center minimize the
144
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
clearly a central point in the interaction.
6.3 GUI Testing
Betweenness is a centrality measure of a vertex or an edge
within a graph [24]. Vertices that occur on many shortest
paths between other vertices have higher betweenness than
those that do not. Similar to vertices betweenness
centrality, edge betweenness centrality is related to
shortest path between two vertices. Edges that occur on
many shortest paths between vertices have higher edge
betweenness.
The reverse engineering approach described in this paper
allows us to extract an abstract GUI behavior
specification.
Our next goal is to perform model-based GUI testing. To
this end, we make use of the QuickCheck Haskell library
tool. QuickCheck is a tool for testing Haskell programs
automatically. The programmer provides a specification of
the program, in the form of properties which functions
should satisfy, and QuickCheck then tests that the
properties hold in a large number of randomly generated
cases. Specifications are expressed in Haskell, using
combinators defined in the QuickCheck library.
QuickCheck provides combinators to define properties,
observe the distribution of test data, and define test data
generators. Considering the application described in the
previous section and its abstract GUI model-based we
could now write some rules and test them through the
QuickCheck tool. To illustrate the approach, we will test if
the application satisfies the following rule: users need to
execute less than three actions to access the main window.
The rule is specified in the Haskell language. From the
windows set we automatically generate randomly cases.
We extract valid GUI sentences from a GUI behavioral
model. Then the rule is tested in a large number of cases
(10000 in this GUI testing process!). The number of
random cases and event lengths are specified by the user.
Each random case is a sequence of valid events associated
with their conditions, actions and the respective window.
In other words, each case is a sequence of possible events,
so all respective conditions are true in this context.
This approach enables to analyze a GUI model using a
model-based testing technique. Though our approach is
non-exhaustive, this is a technique which allows us to test
the quality of models at a lower cost than other exhaustive
techniques such as model checking. This section's focus is
on GUI testing. Coverage criteria for GUIs are important
rules that provide an objective measure of test quality.
We plan to include coverage criteria to help determine
whether a GUI has been adequately tested. These
coverage criteria use event sequences to specify a measure
of test adequacy. Since the total number of permutations
of event and condition sequences in any GUI is extremely
large, the GUI's hierarchical structure must be exploited to
identify the important event sequences to be tested.
Fig. 4 HSM's betweenness values
Figure 4 provides the obtained result when applying the
betweenness algorithm. Betweenness values are expressed
numerically for each vertices and edges. Highest
betweenness edges values are represented by thicker
edges. Some states and edges have the highest
betweenness, meaning they act as a hub from where
different parts of the interface can be reached. Clearly
they represent a central axis in the interaction between
users and the system. In a top down order, this axis
traverses
the
following
states
patStartstate0,
patStartstate1,
startAppstate0,
startAppstate1,
docStartstate0 and docStartstate1. States startAppstate0
and startAppstate1 are the main states of the startApp
window's state machine.
The Main window has the highest betweenness, meaning it
acts as a hub from where different parts of the interface
can be reached. Clearly it will be a central point in the
interaction.
6.4 Conclusions
This Section described the results obtained with the
implemented tool when applying it to a larger interactive
system. The chosen interactive system case study is
145
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
information about the interaction. This means that we are
able to analyze the quality of the user interface, from the
users perspective, without having to resort to external
metrics which would imply testing the system with real
users, with all the costs that process carries.
Additionally, we have explored the possibility of
analyzing the graphs via a testing approach, and how best
to generate test cases. It must be noted that, while the
approach enables us to analyze aspects of user interface
quality without resorting to human test subjects, the goal
is not to replace user testing. Ultimately, only user testing
will provide factual evidence of the usability of a user
interface. The possibility of performing the type of
analysis we are describing, however, will help in gaining a
deeper understanding of a given user interface. This will
promote the identification of potential problems in the
interface, and support the comparison of different
interfaces, complementing and minimizing the need to
resort to user testing. Similarly, while the proposed
metrics and analysis relate to the user interface that can be
inferred from the code, the approach is not proposed as an
alternative to actual code analysis.
Metrics related to the quality of the code are relevant, and
indeed the tool is also able to generate models that capture
information about the code itself. Again, we see the
proposed approach as complementary to that style of
analysis. Results show the reverse engineering approach
adopted is useful but there are still some limitations. One
relates to the focus on event listeners for discrete events.
This means the approach is not able to deal with
continuous media and synchronization/timing constraints
among objects. Another has to due with layout
management issues. The tool cannot extract, for example,
information about overlapping windows since this must be
determined at run time. Thus, we cannot find out in a
static way whether important information for the user
might be obscured by other parts of the interface. A third
issue relates to the fact that generated models reflect what
was programmed as opposed to what was designed.
Hence, if the source code does the wrong thing, static
analysis alone is unlikely to help because it is unable to
know what the intended outcome was. For example, if an
action is intended to insert a result into a text box, but
input is sent to another instead. However, if the design
model is available, the tool can be used to extract a model
of the implemented system, and a comparison between the
two can be carried out.
Additionally, using graph operations, models from
different implementations can be compared in order to
assess whether two systems correspond to the same
design, or to identify differences between versions of the
same system.
related to a healthcare management system (HMS). The
HMS system is implemented in Java/Swing programming
language and implement operations to allow for patients,
doctors and bills management. A description of main
HMS windows has been provided, and tool results have
been described. The tool enabled the extraction of
different behavioral models. Methodologies have been
also applied automating the activities involved in GUI
model-based reasoning, such as, pagerank and
betweenness algorithms. GUI behavioral metrics have
been used as a way to analyze GUI quality. This case
study demonstrated that the tool enables the analysis of
real interactive applications written by third parties.
7. Discussion
The previous section has illustrated how the implemented
tool makes possible high-level graphical representation of
GUI behavior from thousand of lines of code. The process
is mostly automatic, and enables reasoning over the
interactive layer of open source systems. Examples of
some of the analysis that can be carried out were provided.
Other uses of the models include, for example, the
generation of test cases, and/or support for model-based
testing. During the development of the framework, a
particular emphasis was placed on developing tools that
are, as much as possible, language independent. Through
the use of generic programming techniques, the developed
tools aim at being retargetable to different user interface
programming toolkits and languages. At this time, the
framework supports (to varying degrees) the reverse
engineering of Java code, either with the Swing or the
GWT (Google Web Toolkit) toolkits, and of Haskell
code, using the WxHaskell GUI library. Originally the
tool was developed for Java/Swing. The WxHaskell and
GWT retargets have highlighted successes and problems
with the initial approach. The amount adaptation and the
time it took to code are distinct. The adaptation to GWT
was easier because it exploits the same parser. The
adaptation to WxHaskell was more complex as the
programming paradigm is different, i.e. functional. Using
the tool, programmers are able to reason about the
interaction between users and a system at a higher level of
abstraction than that of code. A range of techniques can be
applied on the generated models. They are amenable, for
example, to analysis via model checking [25]. Here
however, we have explored alternative, lighter weight
approaches.
Considering that the graphs generated by the reverse
engineering process are representations of the interaction
between users and system, we have shown how metrics
defined over those graphs can be used to obtain relevant
146
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
methods and functional strategies regarding the reverse
engineering of interactive applications,” in Interactive Systems,
Design, Specifications and Verification, Lecture Notes in
Computer Science. DSV-IS 2006, the XIII International
Workshop on Design, Specification and Verification of
Interactive System, Dublin, Ireland, pp. 137–150, Springer
Berlin / Heidelberg, July 2006.
[2] J. C. Silva, J. C. Campos, and J. Saraiva, “Models for the
reverse engineering of java/swing applications,” ATEM 2006,
3rd International Workshop on Metamodels, Schemas,
Grammars and Ontologies for Reverse Engineering, Genova,
Italy, October 2006.
[3] J. C. Silva, J. C. Campos, and J. Saraiva, “A generic library
for gui reasoning and testing,” in In ACM Symposium on
Applied Computing, pp. 121–128, March 2009.
[4] F. Tip, “A survey of program slicing techniques,” Journal of
Programming Languages, september 1995.
[5] D. Cerri, A. Fuggetta, D. Cerri, A. Fuggetta, and C. P. D.
Milano, “Open standards, open formats, and open source,”
2007.
[6] C. Benson, “Professional usability in open source projects,”
in Conference on Human Factors in Computing Systems, pp.
1083–1084, ACM Press, 2004.
[7] H. Nakajima, T. Masuda, and I. Takahashi, “Gui ferret: Gui
test tool to analyze complex behavior of multi-window
applications,” in Proceedings of the 2013 18th International
Conference on Engineering of Complex Computer Systems,
ICECCS ’13, (Washington, DC, USA), pp. 163–166, IEEE
Computer Society, 2013.
[8] C. E. Silva and J. C. Campos, “Combining static and
dynamic analysis for the reverse engineering of web
applications,” in Proceedings of the 5th ACM SIGCHI
Symposium on Engineering Interactive Computing Systems,
EICS ’13, (New York, NY, USA), pp. 107–112, ACM, 2013.
[9] D. H. Department and D. Horton, “Software testing.”
[10] C. D. Roover, I. Michiels, K. Gybels, and K. Gybels, “An
approach to high-level behavioral program documentation
allowing lightweight verification,” in In Proc. of the 14th IEEE
Int. Conf. on Program Comprehension, pp. 202–211, 2006.
[11] G. Sala¨un, G. Salan, G. Salan, C. Attiogb, C. Attiogb, C.
Attiogb, M. Allem, M. Allem, and M. Allem, “Verification of
integrated specifications using pvs.”
[12] J.-C. Filliatre, “Program verification using coq introduction
to the why tool,” 2005.
[13] ISO, ISO 9241-11: Ergonomic requirements for office work
with visual display terminals (VDTs) – Part 11: Guidance on
Usability. International Organization for Standardization, 1998.
[14] F. A. Fontana and S. Spinelli, “Impact of refactoring on
quality code evaluation,” in In Proceeding of the 4th workshop
on Refactoring toolsi, Honolulu, HI, USA, 2011. ACM.
Workshop held in conjunction with ICSE 2011, 2011.
[15] J. Al Dallal, “Object-oriented class maintainability
prediction using internal quality attributes,” Inf. Softw.
Technol., vol. 55, pp. 2028–2048, Nov. 2013.
[16] ISO/IEC, “Software products evaluation,” 1999. DIS
14598-1.
[17] J. Nielsen, Usability Engineering. San Diego, CA:
Academic Press, 1993.
[18] I. Stamelos, L. Angelis, A. Oikonomou, and G. L. Bleris,
“Code quality analysis in open source software development,”
8. Conclusions
In what concerns interactive open source software
development, two perspectives on quality can be
considered. Users, on the one hand, are typically
interested on what can be called external quality: the
quality of the interaction between users and system.
Programmers, on the other hand, are typically more
focused on the quality attributes of the code being
produced. This work is an approach to bridging this gap
by allowing us to reason about GUI models from source
code. We described GUI models extracted automatically
from the code, and presented a methodology to reason
about the user interface model. A number of metrics over
the graphs representing the user interface were
investigated. An approach to testing the graph against
desirable properties of the interface was also put forward.
A number of issues still needs addressing. In the example
used throughout the paper, only one windows could be
active at any given time (i.e., windows were modal). The
tool is also able to work with non-model windows (i.e.,
with GUIs where users are able to freely move between
open application windows). In that case, however, nodes
in the graph come to represents sets of open windows
instead of a single active window. While all analysis
techniques are still available, this new interpretation of
nodes creates problems in the interpretation of some
metrics that need further consideration. The problem is
exacerbated when multiple windows of a given type are
allowed (e.g., multiple editing windows). Coverage
criteria provide an objective measure of test quality. We
plan to include coverage criteria to help determine
whether a GUI has been adequately tested. These
coverage criteria use events and event sequences to
specify a measure of test adequacy. Since the total number
of permutations of event and condition sequences in any
GUI is extremely large, the GUI's hierarchical structure
must be exploited to identify the important event
sequences to be tested.
This work presents an approach to the analysis of
interactive open source systems through reverse
engineering process. Models enable us to reason about
both metrics of the design, and the quality of the
implementation of that design. Our objective has been to
investigate the feasibility of the approach. We believe this
style of approach can feel a gap between the analysis of
code quality via the use of metrics or other techniques,
and usability analysis performed on a running system with
actual users.
References
[1] J. C. Silva, J. C. Campos, and J. Saraiva, “Combining formal
147
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.
ACSIJ Advances in Computer Science: an International Journal, Vol. 3, Issue 1, No.7 , January 2014
ISSN : 2322-5157
www.ACSIJ.org
Information Systems Journal, vol. 12, pp. 43–60, 2002.
[19] Y. S. Yoon and W. C. Yoon, “Development of quantitative
metrics to support ui designer decision-making in the design
process,” in Human-Computer Interaction. Interaction Design
and Usability, pp. 316–324, Springer Berlin / Heidelberg, 2007.
[20] H. Thimbleby and J. Gow, “Applying graph theory to
interaction design,” pp. 501–519, 2008.
[21] S. P. Jones, J. Hughes, L. Augustsson, et al., “Report on the
programming language haskell 98,” tech. rep., Yale University,
Feb. 1999.
[22] R. Lammel and J. Visser, “A STRAFUNSKI application
letter,” tech. rep., CWI, Vrije Universiteit, Software
Improvement Group, Kruislaan, Amsterdam, 2003.
[23] P. Berkhin, “A survey on pagerank computing,” Internet
Mathematics, vol. 2, pp. 73–120, 2005.
[24] S. Y. Shan and et al., “Fast centrality approximation in
modular networks,” 2009.
[25] J. C. Campos and M. D. Harrison, “Interaction engineering
using the ivy tool,” in ACM Symposium on Engineering
Interactive Computing Systems (EICS 2009), (New York, NY,
USA), pp. 35–44, ACM, 2009.
148
Copyright (c) 2014 Advances in Computer Science: an International Journal. All Rights Reserved.