volume six number two july

Transcription

volume six number two july
VOLUME SIX NUMB ER TWO J UL Y - D ECEMBER 2008
P L AT F O R M
P L AT F O R M
Volume 6 Number 2
2
6
13
21
27
31
38
47
52
65
77
85
91
96
105
111
116
122
129
137
145
152
158
166
VOLUME SIX NU MB ER TWO JU L Y - D EC EMB ER 2008
Mission-Oriented Research: CARBON DIOXIDE MANAGEMENT
Separation Of Nitrogen From Natural Gas By Nano- Porous Membrane Using Capillary Condensation
Farooq Ahmad, Hilmi Mukhtar, Zakaria Man, Binay. K. Dutta
Mission-Oriented Research: DEEPWATER TECHNOLOGY
Recent Developments In Autonomous Underwater Vehicle (AUV) Control Systems
Kamarudin Shehabuddeen, Fakhruldin Mohd Hashim
Mission-Oriented Research: GREEN TECHNOLOGY
Enhancement Of Heat Transfer Of A Liquid Refrigerant In Transition Flow In The Annulus Of A DoubleTube Condenser R. Tiruselvam, Chin Wai Meng, Vijay R Raghavan
Mission-Oriented Research: PETROCHEMICAL CATALYSIS TECHNOLOGY
Fenton And Photo-Fenton Oxidation Of Diisopropanolamine
Abdul Aziz Omar, Putri Nadzrul Faizura Megat Khamaruddin, Raihan Mahirah Ramli
Synthesis Of Well-Defined Iron Nanoparticles On A Spherical Model Support
Noor Asmawati Mohd Zabidi, P. Moodley, P. C. Thüne, J. W. Niemantsverdriet
Technology Platform: FUEL COMBUSTION
Performance And Emission Comparison Of A Direct-Injection (DI) Internal Combustion Engine Using
Hydrogen And Compressed Natural Gas As Fuels
Abdul Rashid Abdul Aziz, M. Adlan A., M. Faisal A. Mutalib
The Effect Of Droplets On Buoyancy In Very Rich Iso-Octane-Air Flames
Shaharin Anwar Sulaiman, Malcolm Lawes
Technology Platform: SYSTEM OPTIMISATION
Anaerobic Co-Digestion Of Kitchen Waste And Sewage Sludge For Producing Biogas
Amirhossein Malakahmad, Noor Ezlin Ahmad Basri, Sharom Md Zain
On-Line At-Risk Behaviour Analysis And Improvement System (E-ARBAIS)
Azmi Mohd Shariff, Tan Sew Keng
Bayesian Inversion Of Proof Pile Test: Monte Carlo Simulation Approach
Indra Sati Hamonangan Harahap, Wong Chun Wah
Element Optimisation Techniques In Multiple DB Bridge Projects
Narayanan Sambu Potty, C. T. Ramanathan
A Simulation Study On Dynamics And Control Of A Refrigerated Gas Plant
Nooryusmiza Yusoff, M. Ramasamy, Suzana Yusup
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
An Interactive Approach To Curve Framing Abas Md Said
Student Industrial Internship Web Portal Aliza Sarlan, Wan Fatimah Wan Ahmad, Dismas Bismo
Hand Gesture Recognition: Sign To Voice System (S2V) Foong Oi Mean, Tan Jung Low, Satrio Wibowo
Parallelization Of Prime Number Generation Using Message Passing Interface
Izzatdin A Aziz, Nazleeni Haron, Low Tan Jung, Wan Rahaya Wan Dagang
Evaluation Of Lossless Image Compression For Ultrasound Images
Boshara M. Arshin, P. A. Venkatachalam, Ahmad Fadzil Mohd Hani
Learning Style Inventory System: A Study To Improve Learning Programming Subject
Saipudnizam Mahamad, Syarifah Bahiyah Rahayu Syed Mansor, Hasiah Mohamed
Performance Measurement – A Balanced Score Card Approach
P. D. D. Dominic, M. Punniyamoorthy, Savita K Sugathan, Noreen I. A.
A Conceptual Framework For Teaching Technical Writing Using 3D Virtual Reality Technology
Shahrina Md Nordin, Suziah Sulaiman, Dayang Rohaya Awang Rambli,
Wan Fatimah Wan Ahmad, Ahmad Kamil Mahmood
Multi-Scale Color Image Enhancement Using Contourlet Transform
Melkamu H. Asmare, Vijanth Sagayan Asirvadam, Lila Iznita
Automated Personality Inventory System Wan Fatimah Wan Ahmad, Aliza Sarlan, Mohd Azizie Sidek
A Fuzzy Neural Based Data Classification System Yong Suet Peng, Luong Trung Tuan
Other Areas
Research In Education: Taking Subjective Based Research Seriously
Sumathi Renganathan, Satirenjit Kaur
Jul - Dec 2008
I SSN 1 5 1 1 - 6 7 9 4
P LA T F OR M
July-December 2008
Advisor: Datuk Dr. Zainal Abidin Haji Kasim
PLATFORM Editorial
Editor-in-Chief:
Prof. Ir. Dr. Ahmad Fadzil Mohd. Hani
Co-Editors:
Assoc. Prof. Dr. Isa Mohd Tan
Assoc. Prof. Dr. Victor Macam Jr.
Assoc. Prof. Dr. Patthi Hussin
Dr. Baharum Baharuddin
Dr. Nor Hisham Hamid
Dr. Shahrina Mohd. Nordin
Subarna Sivapalan
Sub-Editor:
Haslina Noor Hasni
UTP Publication Committee
Chairman: Dr. Puteri Sri Melor
Members:
Prof. Ir. Dr. Ahmad Fadzil Mohamad Hani
Assoc. Prof. Dr. Madzlan Napiah
Assoc. Prof. Dr. M. Azmi Bustam
Dr. Nidal Kamel
Dr. Ismail M. Saaid
Dr. M. Fadzil Hassan
Dr. Rohani Salleh
Rahmat Iskandar Khairul Shazi Shaarani
Shamsina Shaharun
Anas M. Yusof
Haslina Noor Hasni
Roslina Nordin Ali
Secretary:
Mohd. Zairee Shah Mohd. Shah
[email protected]
Address:
PLATFORM Editor-in-Chief
Universiti Teknologi PETRONAS
Bandar Seri Iskandar, 31750 Tronoh
Perak Darul Ridzuan, Malaysia
h ttp : //w w w . u t p . e d u . m y
[email protected]
[email protected]
Telephone +(60)5 368 8239
Facsimile
+(60)5 365 4088
Copyright © 2008
Universiti Teknologi PETRONAS
ISSN 1511-6794
Contents
Mission-Oriented Research: CARBON DIOXIDE MANAGEMENT
Separation Of Nitrogen From Natural Gas By Nano-Porous Membrane Using Capillary
Condensation
Farooq Ahmad, Hilmi Mukhtar, Zakaria Man, Binay. K. Dutta
Mission-Oriented Research: DEEPWATER TECHNOLOGY
Recent Developments In Autonomous Underwater Vehicle (Auv) Control Systems
Kamarudin Shehabuddeen, Fakhruldin Mohd Hashim
Mission-Oriented Research: GREEN TECHNOLOGY
Enhancement Of Heat Transfer Of A Liquid Refrigerant In Transition Flow In The Annulus Of A
Double-Tube Condenser
R. Tiruselvam, Chin Wai Meng, Vijay R Raghavan
Mission-Oriented Research: PETROCHEMICAL CATALYSIS TECHNOLOGY
Fenton And Photo-Fenton Oxidation Of Diisopropanolamine
Abdul Aziz Omar, Putri Nadzrul Faizura Megat Khamaruddin, Raihan Mahirah Ramli
Synthesis Of Well-Defined Iron Nanoparticles On A Spherical Model Support
Noor Asmawati Mohd Zabidi, P. Moodley, P. C. Thüne, J. W. Niemantsverdriet
Technology Platform: FUEL COMBUSTION
Performance And Emission Comparison Of A Direct-Injection (Di) Internal Combustion Engine
Using Hydrogen And Compressed Natural Gas As Fuels
Abdul Rashid Abdul Aziz, M. Adlan A., M. Faisal A. Mutalib
The Effect Of Droplets On Buoyancy In Very Rich Iso-Octane-Air Flames
Shaharin Anwar Sulaiman, Malcolm Lawes
Technology Platform: SYSTEM OPTIMISATION
Anaerobic Co-Digestion Of Kitchen Waste And Sewage Sludge For Producing Biogas
Amirhossein Malakahmad, Noor Ezlin Ahmad Basri, Sharom Md Zain
On-Line At-Risk Behaviour Analysis And Improvement System (E-Arbais)
Azmi Mohd Shariff, Tan Sew Keng
Bayesian Inversion Of Proof Pile Test: Monte Carlo Simulation Approach
Indra Sati Hamonangan Harahap, Wong Chun Wah
Element Optimisation Techniques In Multiple Db Bridge Projects
Narayanan Sambu Potty, C. T. Ramanathan
A Simulation Study On Dynamics And Control Of A Refrigerated Gas Plant
Nooryusmiza Yusoff, M. Ramasamy, Suzana Yusup
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
An Interactive Approach To Curve Framing
Abas Md Said
Student Industrial Internship Web Portal
Aliza Sarlan, Wan Fatimah Wan Ahmad, Dismas Bismo
Hand Gesture Recognition: Sign To Voice System (S2V)
Foong Oi Mean, Tan Jung Low, Satrio Wibowo
Parallelization Of Prime Number Generation Using Message Passing Interface
Izzatdin A Aziz, Nazleeni Haron, Low Tan Jung, Wan Rahaya Wan Dagang
Evaluation Of Lossless Image Compression For Ultrasound Images
Boshara M. Arshin, P. A. Venkatachalam, Ahmad Fadzil Mohd Hani
Learning Style Inventory System: A Study To Improve Learning Programming Subject
Saipudnizam Mahamad, Syarifah Bahiyah Rahayu Syed Mansor, Hasiah Mohamed
Performance Measurement – A Balanced Score Card Approach
P. D. D. Dominic, M. Punniyamoorthy, Savita K Sugathan, Noreen I. A.
A Conceptual Framework For Teaching Technical Writing Using 3d Virtual Reality Technology
Shahrina Md Nordin, Suziah Sulaiman, Dayang Rohaya Awang Rambli,
Wan Fatimah Wan Ahmad, Ahmad Kamil Mahmood
Multi-Scale Color Image Enhancement Using Contourlet Transform
Melkamu H. Asmare, Vijanth Sagayan Asirvadam, Lila Iznita
Automated Personality Inventory System
Wan Fatimah Wan Ahmad, Aliza Sarlan, Mohd Azizie Sidek
A Fuzzy Neural Based Data Classification System
Yong Suet Peng, Luong Trung Tuan
Other Areas
Research In Education: Taking Subjective Based Research Seriously
Sumathi Renganathan, Satirenjit Kaur
VOLUME Six NUMBER two july - december 2008 PLATFORM
2
6
13
21
27
31
38
47
52
65
77
85
91
96
105
111
116
122
129
137
145
152
158
166
1
Mission-Oriented Research: CARBON DIOXIDE MANAGEMENT
Separation of Nitrogen from Natural gas
by Nano-porous Membrane
Using Capillary Condensation
Farooq Ahmad*, Hilmi Mukhtar, Zakaria Man and Binay. K. Dutta,
Universiti Teknologi Petronas, 31750 Tronoh, Perak Darul Ridzuan, Malaysia
*[email protected]
ABSTRACT
In the present work we have explored the potential of a nano-porous membrane to perform the separation job
of binary mixture of methane/nitrogen by capillary condensation. In case of methane/nitrogen permeation
rate up to 700 gmol/m2.s.bar has been achieved at a temperature lower than the critical temperature of the
permeating species and higher than the critical temperature of the non-permeating species. The results have
the potential to be used for further refining and optimising the process conditions to exploit this strategy for
large scale removal of nitrogen from methane at a low cost.
Keywords: capillary condensation, nano-porous membrane, natural gas, nitrogen, permeability
INTRODUCTION
practically an acceptable flux and selectivity could be
achieved by this technique.
Raw natural gas contains many impurities such as, acid
gases (carbon dioxide and hydrogen sulfide), lower
hydrocarbons (propane and butane) and nitrogen.
Studies performed by the Gas research institute reveal
that 14% of known reserves in the United States are
sub-quality due to high nitrogen content [Huggman
et al (1993)]. The conventional cryogenic route is not
favoured as it requires a lot of energy. Gas permeation
through a nano-porous membrane occurs primarily by
Knudsen diffusion although the interaction between
the permeating molecules and the pore wall may cause
other mechanisms to prevail such as surface diffusion
[Jaguste et al. (1995); Uholhorn et al. (1998); Wijaman,
S et al. (1995)]. Multi-layer adsorption occurs and is
followed by capillary condensation. In an earlier paper
[Ahmad et al. (2005)] we have reported an analysis of
separation of lower hydrocarbon from natural gas
by capillary condensation. It was established that
RESULTS AND DISCUSSIONS
The technique presented by Lee and Hwang are
widely used for the description of condensable
gases through small pores of the membranes [Lee
and Hwang (1986)]. They investigated the transport
of condensable vapours through micro-porous
membrane and predict six flow regimes depending
on the pressure distribution and the thickness of the
adsorbed layer. Here we consider the case of complete
filling of a pore, with condensate in both upstream and
downstream. For the condensation to occur in pore
at both upstream and downstream face of membrane
the condensation pressure (Pcon) should be lesser
than both upstream pressure (feed pressure P1 ) and
downstream pressure (Permeate pressure P2) across
the membrane at certain feed temperature greater
This paper was presented at the 15th Regional Symposium on Chemical Engineering
In Conjunction With 22nd Symposium of Malaysian Chemical Engineers, Kuala Lumpur, 2 - 3 December 2008
2
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Mission-Oriented Research: CARBON DIOXIDE MANAGEMENT
than the critical temperature of the methane and
lesser than the critical temperature of carbon dioxide.
This situation is depicted in Figure 1. For this case,
the entire pore is filled with the capillary condensate
and apparent permeability is given by the following
equation [Lee and Hwang (1986)]
Pt =
2
Kd ρ R T
P (r − t ) 2 P2
(r − t )
×[ 2
ln 1
ln ]
µ M ( P1 − P2 )
r
P0 r 2
P0
(1)
where, Kd = ρπr4/8M, t1 and t2 are the thicknesses of
the adsorbed layer at upstream and downstream face
pore respectively. The thickness of the absorbed layer
at upstream and downstream is assumed as 10 times
the molecular diameter of methane. Since methane
is the more permeating component, the selectivity of
methane over nitrogen is given by
α=
xCH4
y CH4
yN 2 xN 2
(2)
where x is the mol fraction in the pore, and y is the
mol fraction in the bulk
The permeability of condensed methane of methane/
nitrogen binary mixture has been calculated using
equation (1) at different temperatures for various pore
sizes. Since the selected pore diameters are small,
condensation occurs at temperature well above the
normal condensation temperature at the prevailing
pressure. This makes the pore flow separation more
attractive than the cryogenic separation process. A wide
range of pore sizes and temperatures were selected for
computation of permeability and separation factors.
The computed results are presented below. Figure 2
gives the permeability of methane with temperature
for different pore sizes and pore lengths equal to
ten times the molecular diameter of methane. With
increasing temperature permeation rate is increased,
because at a higher temperature more pressure is
required to cause capillary condensation inside the
pore. Figure 2 shows that even at moderate pressure
and temperature slightly below °C, an appreciable
permeability can be achieved. The permeation rate
is reduced at low pore size, but at low pore size the
condensation pressure is reduced and we require
lesser pressure at the feed side to cause condensation
inside the pore. Based on solubility of nitrogen in
condensed methane using Peng-Robinson equation of
state, the separation factor binary mixtures methane/
nitrogen has been calculated by using equation (2)
and is shown in Figure 3. From Figure 3 it can be
seen that separation factor decreases with increasing
temperature, because as stated earlier, at a higher
temperature, more feed pressure is required to cause
capillary condensation and thus solubility of nitrogen
in liquid methane increases and thus separation factor
decreases with increasing temperatures. From Figure
3 it is concluded that a reasonable separation factor
Figure 1. Schematic of condensation flow through a nano pore.
900
500
P (gmol / s.m2.bar)
700
600
500
Separation factor
5 nm
10 nm
20 nm
30 nm
40 nm
50 nm
800
400
300
200
4% N2- 96% CH4
8% N2- 92% CH4
12% N2- 88% CH4
16% N2- 84% CH4
400
300
200
100
100
0
100
110
120
130
140
150
160
170
180
0
100
190
T (K)
110
120
130
140
150
160
170
180
T(K)
Figure 2. Effect of pore size on permeability of methane at
different temperatures.
Figure 3. Separation factor of N2/CH4 binary mixtures at various
temperatures.
VOLUME Six NUMBER two july - december 2008 PLATFORM
3
Mission-Oriented Research: CARBON DIOXIDE MANAGEMENT
40
Experimental data
Model Predicted data
600
35
400
300
200
25
20
15
10
100
0
360
Data Predicted by Kelvin equation
Experimental data
30
500
Pcon (Bar)
Separation factor
700
5
380
400
420
T (K)
440
460
0
360
480
380
400
420
T(K)
440
460
480
Figure 4. Comparison of Experimental data for methanol
hydrogen separation by [Sperry et al. (1991)] with Model predicted
data [Ahmad et al. (2008)].
Figure 5. Comparison of Experimental data for methanol
hydrogen separation by [Speery et al. (1991)] with the data
predicted using Kelvin equation [Ahmad et al. (2008)]
can be achieved using a nano-porous membrane by
capillary condensation which justifies that the nanoporous membrane using capillary condensation has a
potential to be used. The model compares reasonably
well with the experimental results which justify the
validity of the model. The comparison is shown in
Figure 4. Figure 4 shows that the separation factor
obtained theoretically is less than the experimentally
calculated separation factor. The reason for this is
that the separation factor is calculated on the basis of
equilibrium solubility of hydrogen in the condensed
phase of methanol. In reality such a system operating
at steady state will be away from equilibrium and
a higher separation factor will be achieved. The
permeability increases with temperature although
the separation factor decreases. A balance should be
struck between the two to decide upon the optimum
operating temperature. For tortuous pores the
increased path length will cause the permeability to
decrease. The comparison between experimental and
predicted values of capillary condensation pressure by
the Kelvin equation given by Sperry et al. is been shown
in Figure 5. The computed results and experimental
data compare well at least up to a temperature of 420 K
(147°C). This establishes the validity of the model up
to a reasonably high temperature for gas separation
applications.
CONCLUSIONS
4
The Kelvin equation was used to calculate the
condition for capillary condensation for predicting
the separation of nitrogen from methane. The
separation factor of methane/nitrogen was analysed
based on the principle that methane will condense
preferentially. High separation factor of up to 439 was
achieved, suggesting that the removal of nitrogen from
natural gas by nano-porous membrane is promising.
Permeation rates were also calculated which are in
agreement for other condensed gas system. Also,
condensation occurred at a temperature much lower
than the normal saturation temperature.
REFERENCES
[1]
R. H. Hugman, E. H. Vidas, P. S. Springer, (1993), ”Chemical
Composition of Discovered and Undiscovered Natural Gas
in the United States”, (Update, GRI-93/0456)
[2]
J. G. Wijmans, R. W. Baker, (1995), “The Solution Diffusion
Model, A review” J. Membr. Sci., (107) 1-21
[3]
R. J. R. Uhlhorn, K. Keizer and A. J. Burggraaf, (1998) “Gas
Separation Mechanism in Micro-porous modified γ-alumina
Membrane”, J. Membr. Sci., (39) 285-300
[4]
D. N. Jaguste and S. K. Bhatia, (1995) ”Combined Surface
and Viscous Flow of Condensable Vapor in Porous Media”,
Chemical Engineering Science, (50) 167-182
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Mission-Oriented Research: CARBON DIOXIDE MANAGEMENT
[5]
F. Ahmad, H. Mukhtar, Z. Man, Binay. K. Dutta, (2007)
“Separation of lower hydrocarbon from Natural gas through
Nano-porous Membrane using Capillary Condensation”,
Chem. Eng. Techno. 30 (9) 1266-1273
[6]
K. H. Lee and Hwang, (1986).The Transport of Condensable
Vapors through a Micro-porous Vycor Glass Membrane J.
Colloid Interface Sci., 110 (2).544-554
[7]
F. Ahmad, H. Mukhtar, Z. Man, Binay. K. Dutta, (2008)
“A theoretical analysis for non-chemical separation of
hydrogen sulfide from Natural gas through Nano-porous
Membrane using Capillary Condensation”, Chem. Eng.
Processing, (47) 2203-2208
Hilmi Mukhtar is currently the Director
of Undergraduate Studies, Universiti
Teknologi PETRONAS (UTP). Before joining
UTP, he served as a faculty member of
Universiti Sains Malaysia (USM) for about 6
years. He was a former Deputy Dean of the
School of Chemical Engineering at USM.
He obtained his BEng in Chemical
Engineering from the University of
Swansea, Wales, United Kingdom in 1990. He completed his MSc
in 1991 and later on his PhD in 1995 from the same university. His
doctoral research focused on the “Characterisation and
Performance of Nanofiltration Membrane”.
He has deep research interests in the area of natural gas
purification using membrane processes. Currently, he is leading
a research project under the Separation & Utilisation of Carbon
Dioxide research group. In this project, the removal of impurities
from natural gas, in particular, carbon dioxide, is the key focus of
the study. In addition, he has research interests in environmental
issues particularly wastewater treatment and carbon trading.
VOLUME Six NUMBER two july - december 2008 PLATFORM
5
Mission-Oriented Research: DEEPWATER TECHNOLOGY
Recent Developments in Autonomous
Underwater Vehicle (AUV) Control Systems
Kamarudin Shehabuddeen* and Fakhruldin Mohd Hashim
Universiti Teknologi PETRONAS, 31750 Tronoh, Perak Darul Ridzuan, Malaysia
*[email protected]
Abstract
Autonomous Underwater Vehicles (AUVs) are tethered free and unmanned. AUVs are powered by onboard energy
sources such as fuel cells and batteries. They are also equipped with devices such as electronic compass, GPS,
sonar sensor, laser ranger, pressure sensor, inclination sensor, roll sensor and controlled by onboard computers to
execute complex preprogrammed missions. In oil and gas sector, two separate categories of AUVs are classified
for application in Exploration and Production (E&P). A survey class vehicle is for inspection of offshore structures
and data acquisition, and a work class vehicle for underwater manipulation and installation. AUV dynamics
involve six degrees of freedom (DOF). Most of the AUV application requires very stringent positioning precision.
However, AUVs’ dynamics are highly nonlinear and the hydrodynamic coefficients of vehicles are difficult to
be accurately estimated due to the variables associated with different operating conditions. Experimental class
AUVs provide an excellent platform for development and testing of various new control methodology and
algorithms to be implemented in developing advanced AUVs. Control performance requirements of an AUV
are most likely to be achieved from control concepts based on nonlinear theory. Recently developed advanced
control methodologies focused on improving the capability of tracking predetermined reference position
and trajectories. Due to increasing depth of operation expected from future AUVs and onboard limited power
supply, the area of future research in AUV control is most likely to expand into incorporating intelligent control
to propulsion system in order to improve power consumption efficiency. This paper presents a survey on some
of the experimental AUVs and the past, recent and future directions of the AUV control methodologies and
technologies.
Keywords: Autonomous Underwater Vehicles (AUV), Experimental AUVs, Past, Recent and Future Control
Methodologies
Introduction
Oceans cover about two-thirds of the whole earth
surface and the living and non-living resources in the
oceans undoubtedly take an important role in human
life. The deep oceans are hazardous particularly due
to the high pressure environment. However, offshore
oil industry is now forced to deal with increasing
depths of offshore oil well. In recent years, oil well
depths (surface of the sea to sea bed) have increased
far beyond the limit of a human diver. This has resulted
in increasing deployment of unmanned underwater
vehicle (UUV) for operations and maintenance of
deepwater oil facilities.
UUV covers both remotely operated vehicles (ROVs)
and autonomous underwater vehicle (AUVs). ROVs
have tethered umbilical cable to enable remote
This paper was presented at the 5th PetroMin Deepwater, Subsea & Underwater Technology, Conference and Exhibition 2007, Kuala Lumpur
29 - 30 October 2007
6
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Mission-Oriented Research: DEEPWATER TECHNOLOGY
The well proven linear controller based on linear
theory may fail to satisfy the control performance
requirements of an AUV. Therefore, interest of many
AUV researchers centred around non-linear control
schemes based on non-linear theory.
2004
1999
1994
1989
1984
1979
1974
1969
1964
Two classes of UUVs are normally used in the E&P
sector. A survey class is for ocean floor mappings,
inspection of offshore structures and data acquisition,
and a work class vehicle for manipulation, installation
and repair operations. Recently, in the UUV market,
ROVs has been gradually replaced by AUVs.
1959
AUVs performing survey, manipulation or inspection
tasks needed to be controlled in six degrees of
freedom [1]. Although the control problem is
kinematically similar to the control of a rigid body in
a six-dimensional space, the effects of hydrodynamic
forces and uncertain hydrodynamic coefficients give
rise to greater challenges in developing a control
system for an AUV. In order for an AUV to achieve
autonomy under the ocean environment, the control
system must have the adaptability and robustness to
the non-linearity and time-variance of AUV dynamics,
unpredictable environmental uncertainties such as
sea current fluctuation and reaction force from the
collision with sea water creatures, modeling difficulty
of hydrodynamic forces.
The offshore oil industry is currently pursuing offshore
oil production with well depths (surface of the sea to
sea bed) that previously would have been considered
technically unfeasible or uneconomical [2]. A study [3],
Figure 1, shows the maximum well depth versus past
years. In 1949, the offshore industry was producing in
about 5 m of water depth and it took 20 years to reach
about 100 m. However, in recent years, the maximum
well depth increased dramatically. It indicates that
the maximum well depth will continue to increase in
the future. The maximum acceptable depth limit for
a human diver is about 300 m. At deeper than these
depths, ocean floor mapping, inspection and repair
operations of facilities must be executed by either
UUVs or inhabited submersibles.
1954
With the technology advancement in battery, fuel
cells, material, computer, artificial intelligence and
communication, AUVs become more popular in
exploring oceanic resources. In the oil and gas
sectors, two separate classes of AUVs are available
for application in Exploration and Production (E&P).
A survey class is for inspection of offshore structures
and data acquisition, and a work class vehicle for
underwater manipulation required for installation of
underwater facilities and equipments.
Market Drivers
1949
AUVs are tethered free, unmanned, powered by
onboard energy sources such as fuel cells and
batteries. AUVs are also equipped with devices such
as electronic compass, GPS, sonar sensor, laser ranger,
pressure sensor, inclination sensor, roll sensor and
controlled by onboard computers to execute complex
preprogrammed missions.
Several recently developed advanced control
methodologies focused on improving the capability
of tracking, given reference position and attitude
trajectories. The objective of the paper is to address
the recent and future directions of AUV control
methodologies.
0
Well Depth (m)
operator to control the operation of the vehicle. Tether
influences the dynamics of vehicle, greatly reducing
maneuverability.
500
1000
1500
2000
2500
Figure 1. Offshore Oil Fields – Maximum well depth versus past
years. Data and figure source [3]
VOLUME Six NUMBER two july - december 2008 PLATFORM
7
Mission-Oriented Research: DEEPWATER TECHNOLOGY
Future of ROVs and AUVs
Due to increasing offshore oil well depth, research
is currently undertaken to enhance the capabilities
of ROVs so that they can become autonomous,
and hence the emergence of AUVs. AUVs are free
from constraints of an umbilical cable and are fully
autonomous underwater robots designed to carry
out specific pre-programmed tasks such as ocean
floor mappings, manipulation, installation and repair
operations of underwater facilities.
PC104+ (Pentium computer CPU 300 MHZ, 128 MB
RAM). The vehicle can be controlled by either an
on-board computer in the autonomous mode or by
operator command using ground station computer
with or without orientation control via tether.
The main challenges in developing high performance
experimental AUVs are the calibration of sensors
and design of appropriate candidate control system,
formulation of control algorithms and implementation
of appropriate actuation amplitudes based on inputs
from sensors.
AUV Control
AUVs performing survey, manipulation or inspection
tasks needed to be controlled in six degrees of
freedom. In order for an AUV to achieve autonomy
under the ocean environment, the control system must
have the adaptability and robustness of non-linearity
and time-variance of AUV dynamics, unpredictable
environmental uncertainties such as sea current
fluctuation and reaction force from the collision with
sea water creatures. Experimental AUVs provide an
excellent platform for the development and testing of
various new control methodologies and algorithms to
be implemented in developing advanced AUVs.
Highlights on Some of the Experimental AUVs
In 1995, with the intention to contribute to AUV
development, the Autonomous Systems Laboratory
(ASL) of the University Of Hawaii has designed and
developed the Omni-directional Intelligent Navigator
(ODIN-I). In 1995, ODIN-I was refurbished and ODINII was born. ODIN-I and ODIN-II have made precious
contribution in the development and testing of various
new control methodology and algorithms.
In 2003, ODIN-III was developed. It has the same
external configuration and major sensors as ODINII, which is a closed-framed spherical shaped vehicle
with eight refurbished thruster assemblies and
a one DOF manipulator [4]. ODIN-III represents
Experimental robotic class AUV. Eight thrusters provide
instantaneous, omni-directional (six DOF) capabilities.
The on-board computer system used in ODIN-III is a
8
A miniature cylindrical shaped AUV called REMUS
(Remote Environmental Monitoring Units) is designed
to conduct underwater scientific experiments and
oceanographic surveys in shallow water [5]. REMUS
is equipped with physical, biological and optical
sensors. A standard REMUS is 19 cm in diameter and
160 cm long. REMUS represents experimental survey
class AUV. The vehicle is controlled by PC104, IBM
compatible computer. REMUS has three motors for
propulsion, yaw control and pitch control. 68HC11
micro-controller is assigned to control the motors.
The communication between the micro-controller
and the CPU is achieved through an RS232 interface.
An even smaller experimental class AUV was developed
to demonstrate that the basic components of an AUV
can be packaged in a 3-inch diameter tube [6]. The
hull was made from standard 3-inch Schedule 40
PVC pipe. The nose cone was machined from a solid
piece of PVC. The tail was made from high density
foam bonded to a piece of aluminum tubing. The
AUV control system is governed by Rabbit 3000 8-bit
microcontroller (RCM3200 module). The processor
runs at 44.2 MHz and is programmed in C language.
AUV Navigation and Sensor Deployments
Inertial navigation system (INS) is primarily used in
AUVs. INS is usually used to measure the angular
displacements of yaw, pitch and roll. A pressure
sensor is used to measure depth. For an experimental
AUV, sonar sensors are used to measure the horizontal
translation displacements. For the commercial scale
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Mission-Oriented Research: DEEPWATER TECHNOLOGY
AUVs, the navigation suite usually consists of an INS, a
velocity sensor and a Global Positioning System (GPS)
receiver. The GPS is used to initialise the navigation
system, and for position determination when the AUV
surfaces at intervals.
tested on Ocean Explorer series AUVs developed by
Florida Atlantic University. The problem with SMC is
the chattering.
Some of the Past and Recent AUV Control
Methodologies
NNC is an adaptive control scheme. NNC poses parallel
information processing features of human brain with
highly interconnected processing elements. The main
attractive features of neural networks include selflearning and distributed memory capabilities [11].
Sliding Mode Control (SMC), Neural Net Control (NNC),
Proportional Integral Derivative (PID) and Adaptive
Controls were among the major control methodologies
explored for the position and trajectory control of an
AUV.
Sliding Mode Control (SMC)
SMC is a non-linear feedback control scheme. Yoerger
1985 [7] 1991 [8] developed SMC methodology
for the trajectory control of AUV. The method can
control non-linear dynamics directly, without need
for linearisation. This attribute is crucial for an AUV,
which exhibits highly non-linear dynamics. The
control methodology only require a single design for
the whole operating range of the vehicle. Therefore,
it is easier to design and implement than a stack of
liberalised controllers.
The method was examined in simulation using a
planar model of the Experimental Autonomous
Vehicle (EAVE) developed by the University of New
Hampshire (UNH).
Yoerger in 1991 [9] developed an adaptive SMC for the
control of experimental underwater vehicle.
Song in 2000 [10] proposed a Sliding Mode Fuzzy
Controller (SMFC). Song determined the parameters
of the SMFC using the method based on Pontryagin’s
maximum principle. Sliding mode control is robust to
the external disturbance and system modeling error.
SMFC takes advantage of the robustness property of
the SMC and interpolating property of the fuzzy in
such a way that the non-linear switching curve could
be estimated and the robustness could be sustained.
The effectiveness of the control philosophy was
Neural Net Control (NNC)
J. Yuh in 1990 [12] proposed a multilayered neural
network for the AUV control, in which the error-back
propagation method is used. The development of
this scheme was motivated by [13] which exhibit that
the teaching error signal, the discrepancy between
the actual output signal and teaching signal can be
approximated by the output error of the control
system.
Tamaki URA in 1990 [14] proposed a Self Organizing
Neural Net Controller System (SONCS) for AUVs.
SONCS had a controller called forward model and
an evaluation mechanism, which were linked with
initiation and modification tools. The dynamics of
the vehicle was represented by a forward model. The
difference between the actual motion of the vehicle
and pre-programmed mission is calculated by an
evaluation mechanism. The fundamental concept
of SONCS uses the backward propagated signals to
adjust the controller. Backward propagated signals
were obtained by the evaluation of the controlled
vehicle’s motion.
K. Ishii in 1995 [15] proposed an on-line adaptation
method “Imaginary Training” to improve the time
consuming adaptation process of the original SONCS.
In this method, SONCS tunes the controller network
through an on-line process in parallel with the actual
operation.
J. S. Wang in 2001 [16] proposed the Neuro-Fuzzy
control systems. Wang investigated the strengths
and weaknesses of the rule formulation algorithms
using the static adaptation and dynamic adaptation
VOLUME Six NUMBER two july - december 2008 PLATFORM
9
Mission-Oriented Research: DEEPWATER TECHNOLOGY
1000
Range (km)
100
Hotel = 10 W
Hotel = 40 W
Hotel = 160 W
10
0.1
1
10
AUV speed (ms-1)
Figure 2. AUV Range as a function of hotel load and speed. Calculations were made for a low drag 2.2m long, 0.6m diameter vehicle with
high efficiency propulsion and batteries providing 3.3 kW.h of energy. Data and figure source: [21]
methods based on clustering techniques to create
the internal structure for the generic types of fuzzy
models.
the system instead of the knowledge of the dynamic
model. The controller was tested on ODIN in the pool
of the University of Hawaii.
Proportional Integral Derivative (PID)
S. Zhao in 2004 [20] proposed an adaptive plus
disturbance observer (DOB) controller. Zhao used
a non-regressor based adaptive controller as an
outer-loop controller and a DOB as an inner-loop
compensator. Zhao carried out experimental work
on the proposed adaptive DOB controller using
ODIN. The experimental work involved determining
the tracking errors in associated predetermined
trajectories. The performance of the proposed
adaptive plus DOB controller was compared with
other controllers such as the PID controller, PID plus
DOB controller and the adaptive controller. The
proposed adaptive DOB controller was reported to
be effective in compensating the errors arising from
external disturbances.
PID is for control over steady state and transient
errors. PID has been widely implemented in process
industries. It is also used as a benchmark against
which any other control scheme is compared [17].
B. Jalving in 1994 [18] proposed three separate PID
technique-based controllers for steering, diving,
and speed control. The roll dynamic was neglected.
Jalving defended it by designing the vertical distance
between the centre of gravity and centre of buoyancy
sufficiently long enough to suppress the moment
from the propeller. The concept was tested on the
Norwegian Defense Research Establishment (NDRE)
AUV. The performance of the flight control system
was reported to be satisfactory during extensive sea
testing.
Adaptive Controls
J. Yuh in 2001 [19] proposed a non-regressor based
adaptive controller. The adaptive law approximated
and revised the parameters defined by the unknown
bounds of the system parameter matrices and then
tunes the controller gains based on performance of
10
Energy Storage, Power Consumption and
Efficiency
Limited onboard power supply gives rise to the
principal design challenge of an AUV system design.
Limited onboard energy storage primarily restrictsthe
range of a vehicle. Key design parameters are specific
energy (energy storage per unit mass) and energy
density (energy per unit volume). Fairly new battery
technologies such as lithium ion and lithium polymer
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Mission-Oriented Research: DEEPWATER TECHNOLOGY
have higher specific energy and energy density than
the conventional lead acid batteries and are now
available for AUV applications.
Two types of power loads are typically identified by
the AUV designers. One is propulsion load (propulsion
and control surface actuator loads) and the other is
hotel load (on-board computers, depth sensor and
navigation sensors).
Propulsion load typically constitutes a large portion of
the power required to operate an AUV. The amount
of energy required to move the vehicle through the
water is a function of both the drag of the vehicle and
efficiency of the propulsion system [21]. Overall loss
of efficiency of the propulsion system is contributed
by electric motor efficiency losses, hydrodynamic
inefficiency of the propeller, shaft seal frictional losses
and viscous losses.
Reduction in hotel load is obviously beneficial,
and is aided by continuing advances in low power
consumption computers and electronic components.
As the hotel power consumption decreases, the
vehicle speed for best range also decreases. This was
shown by a study on an existing AUV [21], in Figure
2. Currently, this research has found only one source
of reference on the relationship between hotel load,
range and speed.
Conclusion
Recently developed advanced control methodologies
focused on improving the capability of tracking predetermined reference positions and trajectories.
Control performance requirements of an AUV have
been achieved from control concepts based on nonlinear adaptive theory rather than the linear theory
based on fixed parameters. Due to the non-linear (time
variant) hydrodynamic coefficients and unforeseen
disturbances that the AUV control has to deal with,
future research work on AUV position and trajectory
control is likely to investigate the effectiveness of
newly developed non-linear DOB based control
concepts.
Due to the increasing oil well depth, future control
methodologies of AUVs will most likely involve
broadening the scope of control from the sole position
and trajectory control method to that of incorporating
intelligence system for power efficient orientated
mission planning, and will also involve intelligent
control based on actuation of control surface and
thrusters to maximise efficiency in consumption of
the limited onboard power supply.
Due to increasing depth of operations expected
from future AUVs and the varying water pressures
and densities with depth, a key area for future
research and development in AUV control is likely to
investigate the possibility of incorporating innovative
power train technologies (between the electric motor
and propeller) with intelligent control in order that
the limited onboard power supply is consumed with
maximum efficiency. Variable propeller pitch with
intelligent control is also likely to optimise power
consumption.
Due to the deep ocean environment where a
human diver could not reach, future control
methodologies of work class AUVs are also likely to
focus on autonomous coordination based control of
cooperating manipulators or humanoid end effectors
between multiple AUVs performing underwater
equipment installation or repair tasks. While the
adaptive control concepts are better suited for
position control of AUVs, the learning capability of
the NNC system may be considered for coordination
of manipulators. I f the newly developed coordination
based control methodologies were to be represented
in a virtual reality environment, it would be easier to
evaluate responsiveness and behavior of cooperating
manipulators.
References
[1]
Gianluca Antonelli, Stefano Chiaverini, Nilajan Sarkar,
and Michael West, “Adaptive Control of an Autonomous
Underwater Vehicle: Experimental Results on ODIN”, IEEE
Transaction on Control Systems Technology, Vol. 9, No. 5,
pp. 756–765, September 2001
[2]
Loius L. Whitcomb, “Underwater Robotics: Out of the
Research Laboratory and Into the Field”, International
Conference on Robotics and Automation, IEEE 2000
VOLUME Six NUMBER two july - december 2008 PLATFORM
11
Mission-Oriented Research: DEEPWATER TECHNOLOGY
[3]
D. Harbinson and J. Westwood, “Deepwater Oil & Gas – An
Overview Of The World Market”, Deep Ocean Technology
Conference, New Orleans, 1998
[18] Bjvrn Jalving, “The ADRE-AUV Flight Control System”, IEEE
Journal of Ocean Engineering, Vol. 19, No. 4, pp. 497–501,
October 1994
[4]
H. T. Choi, A. Hanai, S. K. Choi, and J. Yuh, “Development of
an Underwater Robot, ODIN-III”, Proceedings of the 2003
IEEE/RSJ International Conference on Intelligent Robots and
Systems, pp. 836–841, Las Vegas, Nevada, October 2003
[5]
Mike Purcell, Chris von Alt, Ben Allen, Tom Austin, Ned
Forrester, Rob Goldsborough and Roger Stokey, “Nee
Capablities of the REMUS Autonomous Under Vehicle”, IEEE
2000
[19] J. Yuh, Michael E. West, P. M. Lee, “An Autonomous
Underwater Vehicle Control with a Non-regressor Based
Algorithm +”, Proceedings of the 2001 IEEE International
Conference on Robotics and Automation, pp. 2363–2368,
Seoul, Korea, May 2001
[6]
Aditya S. Gadre, Jared J. Mach, Daniel J. Stilwell, Carl E. Eick,
“Design of a Prototype Miniature Autonomous Underwater
Vehicle”, Proceedings of the 2003 IEEE/RSJ International
Conference on Intelligent Robots and Systems, pp. 842–
846, Las Vegas, Nevada, October 2003
[7]
D. Yoerger, J. Newman, “Demonstration of closed-loop
Trajectory Control of an Underwater Vehicle”, 1985
Proceedings of OCEANS, vol 17, pp. 1028–1033, Nov. 1985
[8]
D. Yoerger, J. Slotine, “Robust Trajectory Control of
Underwater Vehicles”, IEEE Journal of Oceanic Engineering,
Vol. OE-10, No.4, pp. 462 – 470, October 1985
[9]
D. Yoerger, J. Slotine, “Adaptive Sliding Control of and
Experimental Underwater Vehicles”, Proceedings of the 1991
IEEE International Conference on Robotics and Automation,
pp. 2746 – 2751, Sacramento, California, April 1991
[10] F. Song and S. Smith, “Design of Sliding Mode Fuzzy
Controllers for Autonomous Underwater Vehicle without
System Model”, OCEANS’2000 IEEE/MTS, pp. 835-840, 2000
[11] Arthur G. O. Mutambara, “Design And Analysis of Control
Systems”, CRC Press 1999
[12] J. Yuh, “A Neural Net Controller for Underwater Robotic
Vehicles,” IEEE Journal of Ocean Engineering, Vol. 15, No. 3,
pp. 161–166, July 1990
[13] J. Yuh, R. Lakshmi, S. J. Lee, and J. Oh, “An Adaptive NeuralNet Controller for Robotic Manipulators”, in Robotics and
Manufacturing, M. Jamshidi and M. Saif, Eds. New York:
ASME, 1990
[14] Tamaki URA, Teruo FUJII, Yoshiaki Nose and Yoji Kuroda,
“Self-Organizing Control System for Underwater Vehicles”,
IEEE 1990
[15] Kazuo Ishii, Teruo Fujii, and Tamaki Ura, “An On-line
Adaptation Method in a Neural Network Based Control
System for AUVs”, IEEE Journal of Ocean Engineering, Vol.
20, No. 3, pp. 221–228, July 1995
[16] Jeen-Shing Wang and C. S. George Lee, “Efficient NeuroFuzzy Control Systems for Autonomous Underwater
Vehicle Control”, Proceedings of the 2001 IEEE International
Conference on Robotics and Automation, pp. 2986–2991
Seoul, Korea, 2001
[20] S. Zhao, J. Yuh, and S. K. Choi, “Adaptive DOB Control
for AUVs”, Proceedings of the 2004 IEEE International
Conference on Robotics and Automation, pp. 4899–4904,
New Orleans, LA, April 2004
[21] James Bellingham, MIT Sea Grant, Cambridge, MA, USA.
(doi:10.1006/rwos.2001.0303)
Kamarudin Shehabuddeen received the
B. Eng. degree in mechanical engineering
from University of Sunderland, UK, in 1996,
and M.S. degree in engineering design
from Loughborough University, UK, in
1998. He is registered with the Board of
Engineers Malaysia (BEM) as a Professional
Engineer (P.Eng.) in the mechanical
engineering discipline.
He worked in Wembley I.B.A.E. Sdn. Bhd., Shah Alam, Malaysia, as
a mechanical design engineer from 1996 – 1997. After receiving
the M.S. degree, he worked with Sanyco Grand Sdn Bhd., Shah
Alam, Malaysia, as a test rig design engineer for automotive brake
master pumps. In 1999, he joined Universiti Teknologi PETRONAS
(UTP), Malaysia, where he is currently a lecturer in the department
of mechanical engineering. His current research interests include
neuro-fuzzy based control, adaptive controls, Global Positioning
System (GPS) based navigation of autonomous underwater
vehicle and unmanned air vehicle. Currently he is pursuing his
PhD degree in mechanical engineering at UTP.
Fakhruldin Mohd. Hashim is currently an
Associate Professor at the Department of
Mechanical
Engineering,
Universiti
Teknologi PETRONAS (UTP). Formerly the
head of the department and now the
Deepwater Technology R&D cluster leader,
he holds a BEng in Mechanical Engineering
(RMIT), an MSc (Eng) in Advanced
Manufacturing Systems & Technology
(Liverpool) and a PhD in Computer Aided Engineering (Leeds).
Currently he is a UTP Senate and Research Strategy Committee
member, and the Chairman of the Research Proposal Evaluation
Committee for UTP. Dr. Fakhruldin’s areas of interest include
Engineering Systems Design, Sub-sea Facilities and Underwater
Robotics. He has been engaged as a consultant in over 20
engineering projects and has secured a number of major R&D
grants during his 20-year career as an academician. He is one of
the assessors for the Malaysian Qualifications Agency (MQA) and
an external examiner of one of the local public university. He is
also a trained facilitator in strategic planning and organisational
management.
[17] Ahmad M. Ibrahim, “Fuzzy Logic for Embedded Systems
Applications”, Elsevier Science 2004
12
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Mission-Oriented Research: GREEN TECHNOLOGY
ENHANCEMENT OF HEAT TRANSFER OF A
LIQUID REFRIGERANT IN TRANSITION FLOW
IN THE ANNULUS OF A DOUBLE-TUBE CONDENSER
R. Tiruselvam1, W. M. Chin1 and Vijay R. Raghavan*
*Universiti Teknologi PETRONAS, 31750 Tronoh, Perak Darul Ridzuan, Malaysia
1OYL R&D Centre, Selangor, Malaysia
Abstract
The basis of the present study is that augmentation can reduce the temperature difference across the condenser
and evaporator of a refrigeration system and increase its energy efficiency. This research is conducted on the
inner tube having a 3D corrugated outer surface. The annulus-side coefficients are determined using the Wilson
Plot technique. It was found that the form of correlation by Hausen for transitional annular flow region with
adjusted empirical constants predicts to a good accuracy the subcooled transitional flow of liquid. For the single
phase heat transfer correlation proposed, all predicted data lie within +5% of the experimental values.
Keywords: Double-Tube Condenser, Transition Flow in Annulus, Augmented Surface, Heat Transfer Enhancement.
Introduction
Condensation heat transfer, both inside and outside
horizontal tubes, plays a key role in refrigeration, airconditioning and heat pump applications. In recent
years, the adoption of substitute working fluids and
new enhanced surfaces for heat exchangers has
thrown up new challenges in condensation heat
transfer research. Well-known and widely established
correlations to predict heat transfer during
condensation may prove to be inaccurate in some
new applications, and consequently a renewed effort
is now being dedicated to the characterization of flow
conditions and associated predictive procedures of
heat transfer with condensing vapour. Much research
effort has been directed at miniaturizing thermal
systems and identifying innovative techniques
for heat transfer enhancement. These techniques
are classified as: passive enhancement techniques
and active enhancement techniques. The double-
tube condenser is an example of the use of passive
enhancement technique. For the given arrangement
of double tube, the refrigerant flows in the annulus
and the cooling water flows in the inner tube, shown
in Figure 1.
Figure 1. Concentric Tube Configuration (Double Tube Heat
Exchanger)
This paper was presented at the 5th European Thermal Sciences Conference, Netherlands
18 - 22 May 2008
VOLUME Six NUMBER two july - december 2008 PLATFORM
13
Mission-Oriented Research: GREEN TECHNOLOGY
Objectives
The purpose of the study is to obtain the single
phase heat transfer coefficient for the subcooled
liquid refrigerant flow in the annulus of the double
tube exchanger. This research is conducted for a 3D
corrugated outer surface enhancement (Turbo-C
Copper Top Cross) on the inner tube used in a doubletube design. It is difficult to generate a mathematical
model for such a complex geometry, and no standard
correlation for the heat transfer coefficient is available
for the Turbo-C copper top cross. Modification
of equations from previous literature will require
extensive measurements of pressure and temperature
of the refrigerant and water sides, and the surface
temperature of the fin and fin root surfaces. With the
above mentioned motivation, the current research
aims to characterize the heat transfer performance
of the Turbo-C Copper Top Cross tube for subcooled
liquid and the enhancement in comparison with a
plain tube. Experiments are conducted to obtain the
necessary data using R-22 (Chlodifluoromethane) as
test fluid and plain tube and Copper Top Cross as test
surfaces.
Experiments
The commissioning of the test section was carried
out using R-22 refrigerant in the annulus and water
as coolant in the inner tube. The primary objective of
these tests is to determine the correlation for overall
heat transfer of the condensation process. The overall
energy flow in the test section can be determined
using three independent routes. These routes use:• temperature increase in the coolant flow
• the mass of condensate collected in the test
section
• circumferentially averaged temperature drop
across the tube wall
Deans et al. (1999) have reported that the maximum
difference in the calculated overall energy flows
using these three routes to analyse the condensation
process was less than 5%. The temperature increase
in the coolant flow is chosen in the present case due
to its ease in experimental measurements as well as
14
in the Wilson Plot. The test facility was designed and
assembled so as to cater for this need. The test facility
is capable of testing either plain or enhanced straight
double-tube condenser at conditions typical of a
vapour compression refrigeration system.
Setup
The double-tube heat exchanger configuration used
in this study is a one-pass, single-track system. Singletrack system means that that only one test section
and one refrigerant compressor may be installed in
the system for individual testing. The compressed
refrigerant is initially condensed into subcooled liquid
using a pre-condenser before it passes through the
double-tube condenser only once as it travels from
the high-side (compressor discharge line) to the lowside (compressor suction line), as shown in Figure
2. The single-track and single-pass system makes it
possible to obtain one successful data point for every
setting. If the operating conditions such as refrigerant
mass flow rate or compressor discharge pressure
are varied (non-geometric variables), it is possible to
obtain additional data points without changing the
test section or compressor. The use of the electronic
expansion valve (EXV) permits such operation. Use
of a conventional expansion device such as the
capillary tube will involve repetition of work where
the refrigerant circuit has to be vacuumed, leak tested
and recharged for the individual length of capillary
tube needed to throttle the refrigerant flow. The
two main variable components in this test facility
are the test section and the refrigerant compressor.
The facility is designed and installed with valves
and fittings for both the refrigerant medium and
the cooling water. This is to allow for quick and easy
replacement of the test section and/or refrigerant
compressor. Each refrigerant compressor has an
individual range of refrigerant flow rate, depending
on the amount of refrigerant charge and compressor
suction and discharge pressure. Three different
refrigerant compressors were chosen (1HP, 2HP, and
3HP) to provide a sufficient range of refrigerant mass
flow rates. A straight horizontal test section was used
in this study, with two types of inner tube, i.e. Plain
Tube (Plain Annulus) and Turbo-C (Enhanced Annulus),
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Mission-Oriented Research: GREEN TECHNOLOGY
Figure 2. Schematic Diagram of Experimental facility
VOLUME Six NUMBER two july - december 2008 PLATFORM
15
Mission-Oriented Research: GREEN TECHNOLOGY
Table 1. Description of the Test Section Inner Tubes
Description
Plain Tube
Turbo-C
Inner Tube Outer Diameter, do
22.2 mm
22.2 mm
Inner Tube Inner Diameter, di
17.82 mm
17.82 mm
Length
2980 mm
2980 mm
Outer Surface
Smooth Surface
3-D Integral Fin
Inner Surface
Smooth Surface
Smooth Surface
Other Information
N.A.
Pitch of Fin = 0.75 mm
Pitch of Corrugation = 5 mm
Fin height = 0.8 mm
with data as shown in Table 1. An illustration of the
enhanced annulus is shown in Figure 3.
Data Reduction & Experimental Uncertainties
In the data run the subcooled liquid from the precooler enters and exits in the same phase. The heat
transferred to the cooling water is kept sufficiently low
to allow the subcooled liquid temperature at the test
section exit to be below the saturation temperature;
hence no phase change occurs. This will allow us to
obtain the single phase heat transfer coefficient for
subcooled liquid.
Q SC = M R ( ∆h) R = M C (Cp )C (∆T )C = U SC (AC )o ( LMTD )
(1)
(2)
1
1
1
=
+ RW +
U SC (AC )o hSC (AC )o
hC (AC )i From equation (1) and (2) the heat transfer coefficient
for single phase liquid is calculated using the Wilson
Plot technique by subtracting the thermal resistance
of the tube wall and cooling water from the overall
thermal resistance. This gives us the heat transfer
coefficient for subcooled liquid given in equation (3).
hSC
 (A ) LMTD 
 (AC )o
 − ((AC )o RW )− 
=  C o
 (A ) h
QSC

 C i C

−1
 

 (3)
Figure 3. Illustration of enhanced surface (Turbo-C)
RW
 kC
hC = 
 Di

 fC 
 Re C Pr C


 2 


0 .5


0 . 63
 f 
 1 . 07 +  900  − 
 + 12 . 7  2  (Pr C − 1 )(Pr C
 Re 
(
)
1
10
Pr
+



C
C

 

(4)
)
−1
3






(5)
Uncertainties in the experimental data were calculated
based on the propagation of error method, described
by Kline and McClintock (1953). Accuracy for various
measurement devices, refrigerant properties and
water properties are given in Table 3 and 4.
The heat transfer coefficient hi for cooling water in
the tube is got from Petukhov and Popov(1963), for
the application range, 0.5 < Pr < 106 and 4000 < Re <
5x106, with a reported uncertainty of 8%.
16
d 
ln  O 
 di 
=
2π k C u L PLATFORM VOLUME Six NUMBER TWO july - december 2008
Mission-Oriented Research: GREEN TECHNOLOGY
Table 2. Uncertainties of Various Measurement Devices
Parameter (Make, Type)
Uncertainties
Water Volume Flow (YOKOGAWA, Magnetic Flow Meter)
+ 0.08% of reading
Refrigerant Mass Flow Meter (YOKOGAWA, Coriolis)
+ 0.05% of reading
Refrigerant Pressure (HAWK, Pressure Transducer)
+ 0.18 psig
Water temperature (CHINO, Pt-100 RTD)
+ 0.1 °C
Refrigerant Temperature (VANCO, Type-T Thermocouple)
+ 0.6 °C
Data provided by OYL R&D Centre Malaysia
Table 3. Uncertainties of Properties
Predicted Properties (R-22)
Uncertainties
Source
Density
+ 0.1%
Isobaric Heat Capacity
+ 1.0%
Viscosity
+ 2.1%
Klein et al. (1997)
Thermal Conductivity
+ 3.7%
McLinden et al. (2000)
Predicted Properties (Water)
Uncertainties
Density
+ 0.02%
Isobaric Heat Capacity
+ 0.3%
Viscosity
+ 0.5%
Thermal Conductivity
+ 0.5%
Predicted Properties (Copper)
Thermal Conductivity
Kamei et al. (1995)
Wagner and Pruß (2002)
Kestin et al. (1984)
Uncertainties
+ 0.5%
Touloukian et al. (1970)
Property data obtained from ASHRAE (2001)
Table 4. Uncertainty Analysis for Experimental Data
Test Sequence
Plain Tube, hsc (W/m2.K)
Turbo-C, hsc (W/m2.K)
Compressor
Highest
Lowest
Highest
Lowest
1 HP
+ 5.40%
+ 4.81%
+ 5.32%
+ 5.04%
2 HP
+ 4.90%
+ 3.86%
+ 4.24%
+ 3.95%
3 HP
+ 4.42%
+3.72%
+ 3.99%
+ 3.89%
VOLUME Six NUMBER two july - december 2008 PLATFORM
17
Mission-Oriented Research: GREEN TECHNOLOGY
Uncertainties in the single phase heat transfer
coefficient (subcooled liquid) are calculated for various
test runs in the smooth and enhanced annulus as a
root-sum-square (RSS) method. Experimental results
and the associated uncertainties are listed in Table 6.
The uncertainties are dominated by the uncertainties
associated with the refrigerant and water properties.
Higher uncertainties were found at higher refrigerant
mass flow rate.
Results & Analysis
reduction and the correlations were evaluated using
property tables in ASHRAE (2001). All subsequent
analysis of correlations is given in non-dimensional
form, shown in Figure 4. Comprehensive reviews of
literature led to selection of a few correlations which
best represent the annulus geometries and flow
characteristics of the fluids in the test section. The
selected correlation was compared with experimental
work on the plain tube. This is done to standardise the
apparatus and assess the applicability of the selected
correlations.
This experimental work was conducted to develop
heat transfer correlations so that the results of the
study could be directly used by the HVAC community
for design and development of double tube condenser
systems. The first step in this effort was comparison
of the current data with available correlations. The
correlations available in the literature range from
purely theoretical ones to those, purely empirical
in nature. All fluid properties required in the data
The one that most accurately represents the range
and conditions of interest was used as the starting
point. These correlations give the basic representation
of the average heat transfer coefficients for a given
set of parametric conditions. Next, it was assumed
that the presence of the fluid inlet and exit fittings
(both refrigerant and cooling water) and surface
enhancement did not alter the form of the heat
transfer coefficient substantially and that any
Figure 4. Comparison of augmented single phase Nusselt number
18
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Mission-Oriented Research: GREEN TECHNOLOGY
difference present can be handled by adjusting the
empirical constants. Base on the above conditions,
the correlation by Hausen (1934) was found to be most
suitable for the following analysis where it applies for
transitional flow development until fully developed
turbulent flow, given in equation (6). The use of Hausen
(1934)’s correlation was reviewed and presented by
Knudsen and Katz (1950) and by Knudsen et al. (1999).
 D
Nu = C (Re) 3 (Pr ) 3 1 +  i
  DO
2
1
2



3



for 2,000 < Re < 35,000
(6)
Considering that the Reynolds Number for the test is
<10,000 the flow of subcooled liquid in both the plain
and enhanced annulus is taken to be in the transition
region. Equation (6) was used to evaluate the Nusselt
type correlation which is given by Hausen (1934).
Thus, the single phase heat transfer coefficient for
subcooled liquid for the plain annulus is:



(7)
Equation (6) was also used to evaluate the Nusselt
type correlation for enhanced annulus since
applicable correlations for transition flow in the
enhanced annulus are lacking. Thus, the single phase
heat transfer coefficient for subcooled liquid flow for
enhanced annulus is:
(Nu)
PLAIN TUBE
LIQUID
( Nu )
TURBO − C
LIQUID
= 0.0055(Re h )
0.8058
= 0.0086 (Re h )
0.8175
 
( PrR) 3 1 +  Di
  DO
1
(PrR)
1
3
 D
1 +  i
  DO






2
3



(8)
2
3
Comparison of experimental Nusselt value using (7)
and (8) against predicted Nusselt value is illustrated in
Figure 5.
Overview Remarks
The overall objective of the present study was to
develop the single phase heat transfer coefficient for
subcooled liquid in transition flow. The correlation
by Hausen (1934) was used for both plain annulus
Figure 5. Experimental Nu vs. Predicted Nu for Subcooled Liquid
and enhanced annulus for the transition region. The
new empirical constants resulted in good prediction
for transitional subcooled liquid flow in the annulus
for both plain annulus and enhanced annulus. By
examining the accuracy of the single phase heat
transfer correlation proposed, all predicted data is
within the +5% of experimental value. The subcooled
liquid flow of plain annulus has an absolute deviation of
+2.52%. Similar results are observed for the subcooled
liquid flow for enhanced annulus with +2.46%.
Nomenclature
C
Coefficient in heat transfer correlation
(∆h)R
Enthalpy change of liquid R-22 (J/kg)
LMTD
Log mean temperature difference
Reh
Reynolds Number of liquid R-22 based on annulus
hydraulic diameter
U
Average Overall heat transfer coefficient (W/m2K)
di
Inner tube inner diameter
do
Inner tube outer diameter
Di
Outer tube inner diameter
Subscripts:
c
Cooling Water
co
copper
R
Refrigerant R-22
SC
subcooled liquid
w
wall
Acknowledgement
The authors wish to acknowledge the support provided for this
research by OYL Research and Development Sdn. Bhd.
VOLUME Six NUMBER two july - december 2008 PLATFORM
19
Mission-Oriented Research: GREEN TECHNOLOGY
References
[1]
ASHRAE, 2001, “Fundamentals Handbook”, Appendix E
Thermophysics Properties
[2]
Deans, J., Sinderman, A., Morrison, J.N.A., 1999, “Use Of
The Wilson Plot Method To Design and Commission A
Condensation Heat Transfer Test Facility,” Two-Phase Flow
Modelling and Experimentation, Edizioni ETS, pp. 351-357
[3]
Hausen, 1934, “C.H., Z. Ver. Dtsch. Ing. Beih.” Verfahrenstech.,
Vol. 91, No. 4
[4]
Kamei, A., Beyerlein, S. W., and Jacobsen, R.T., 1995,
“Application of Nonlinear Regression in the Development
of a Wide Range Formulation for HCFC-22,” International
Journal of Thermophysics, Vol. 16, No. 5, pp. 1155-1164
[5]
Kestin, J., Sengers, J. V., Kamgar-Parsi, B., and Levelt Sengers,
J.M.H., 1984, “Thermo Physical Properties of Fluid H2O,”
Journal of Physical and Chemical Reference Data, Vol. 13,
No. 1, pp. 175-183
[6]
Klein, S. A., McLinden, M. O., 1997, “An Improved Extended
Corresponding States Method for Estimation of Viscosity
of Pure Refrigerants and Mixtures,” International Journal of
Refrigeration, Vol. 20, No. 3, pp. 208-217
[7]
Kline, S., and McClintok, F., 1953, “Describing Uncertainties
in Single-Sample Experiments,” Mechanical Engineering,
Vol. 75, pp. 3-8
[8]
Knudsen, J. G., and Katz, D. L., 1950 “Chemical Engineering
Progress”, Vol. 46, pp. 490
[9]
Knudsen, J. G., Hottel, H. C., Sarofim, A. F., Wankat, P. C.,
Knaebel, K. S., 1999, “Heat and Mass Transfer”, Ch. 5, McGrawHill, New York.
[10] Mclinden, M. O., Klein, S. A., Perkins, R. A., 2000, “An
Improved Extended Corresponding States Model of Thermal
Conductivity Refrigerants and Refrigerant Mixtures,”
International Journal of Refrigeration, Vol. 23, pp. 43-63
[11] Petukhov, B. S., and Popov, V. N., 1963 “Theoretical
Calculation of Heat Exchanger in Turbulent Flow in Tubes of
an Incompressible Fluid with Variable Physical Properties,”
High Temp., (1/1), pp. 69-83
[12] Tiruselvam, R., 2007, “Condensation in the Annulus of
a Condenser with an Enhanced Inner Tube”, M.S. thesis
(research), University Tun Hussein Onn Malaysia
[15] Wagner, W., and Pruß, A., 2002, “New International
Formulation for the Thermodynamic Properties of Ordinary
Water Substance for General and Scientific Use,” Journal of
Physical and Chemical Reference Data, Vol. 31, No. 2, pp.
387-535
R. Tiruselvam is currently a PhD candidate
at the Faculty of Mechanical Engineering,
Universiti Teknologi Petronas.
He
received his Master of Engineering
(Research) in Mechanical Engineering from
University Tun Hussein Onn Malaysia
(UTHM) in 2007. Previously, he was
conferred the B. Eng. (Hons) (Mechanical)
from UTHM and Dip. Mechanical Eng. from
POLIMAS. His research interest is mainly in heat transfer
enhancement in thermal systems. He has been collaborating with
OYL R&D Centre for a period of 4 years. Currently he holds a
position as Research Engineer in OYL R&D Centre.
Chin Wai Meng is the Research Manager of
OYL Research & Development Centre Sdn
Bhd, the research and design centre for
OYL Group, whose primary business is in
the Heating, Ventilation and AirConditioning (HVAC). He has been with the
company for the past 19 years where his
expertise is in the testing of air-conditioning
units to determine performance and
reliability. He also has experience in the design and construction
of psychrometric test facilities. For the past 3 years, he has
established and led the Research Department which specialises in
the research on Heat Transfer and Refrigeration Systems. Mr. Chin
holds a Bachelor’s degree in Mechanical Engineering from
Universiti Malaya and he is currently pursuing a Master of Science
in Universiti Teknologi Petronas.
Vijay R. Raghavan is a professor of
Mechanical Engineering at Universiti
Teknologi Petronas. Earlier he was a
professor of Mechanical Engineering at
Universiti Teknologi Tun Hussein Onn
Malaysia (UTHM) and at the Indian Institute
of Technology Madras. His areas of interest
are Thermofluids and Energy. He obtained
his PhD in Mechanical Engineering in the
year 1980 from the Indian Institute of Technology. In addition to
teaching and research, he is an active consultant for industries in
Research and Development, Design and Troubleshooting.
[13] Tiruselvam, R., Vijay R. Raghavan., Mohd. Zainal B. Md.
Yusof., 2007, “Refrigeration Efficiency Improvement Via
Heat Transfer Enhancement”, Paper No. 030, Engineering
Conference “EnCon”, Kuching, Sarawak, Malaysia, Dec. 2729
[14] Toulaukian, Y. S., Powell, R. W., HO, C. Y., and Klemens, P. G.,
1970, “Thermophysical Properties of Matter, Vol. 1, Thermal
Conductivity, Metallic Elements and Alloys,” IFI/Plenum,
New York
20
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Mission-Oriented Research: pETROCHEMICAL CATALYSIS TECHNOLOGY
Fenton and photo-Fenton Oxidation of
Diisopropanolamine
Abdul Aziz Omar*, Putri Nadzrul Faizura Megat Khamaruddin and Raihan Mahirah Ramli
Universiti Teknologi PETRONAS, 31750 Tronoh, Perak Darul Ridzuan, Malaysia.
*[email protected]
ABSTRACT
Diisopropanolamine has been widely used as an additive in cosmetic and personal care products, metalworking
fluids as a corrosion inhibitor and lubricity agent, pharmaceutical industry for drug intermediates as well as solvent
to remove acid gas from raw natural gas. Although it is well applied in industry, the in-situ wastewater treatment
method for diisopropanolamine contaminated wastewater has not yet been developed. The applicability of
Fenton’s reagent and photo-Fenton for the degradation of diisopropanolamine was investigated. The effect
of H2O2 concentration towards the degradation was investigated by changing the concentration of H2O2 while
keeping the initial COD, concentration of FeSO4 , pH and temperature constant at 50,000 mg/L, 4.98 g/L, 3 and
30 °C respectively. 31% and 24% of the diisopropanolamine degradation were achieved for Fenton and photoFenton respectively at Fe:H2O2 ratio 1:50. Further research work will be conducted to increase the degradation
efficiency and determine other optimum parameters.
Keywords: Diisopropanolamine, wastewater, fenton oxidation, photo-fenton oxidation
Introduction
Diisopropanolamine (DIPA) is a secondary amine of
aliphatic amine group. It has been widely used as
an additive in cosmetic and personal care products,
metalworking fluids as a corrosion inhibitor and
lubricity agent, pharmaceutical industry for drug
intermediates as well as solvent to remove acid gas
from raw natural gas. Wastewater contaminated with
DIPA has an extremely high chemical oxygen demand
(COD) which exceeds the limits set by local authorities.
High influent COD levels make the biological treatment
of the wastewater not possible. Typical practice is to
store the wastewater in a buffer tank prior to pick up
by a licensed scheduled waste contractor. The cost
of the wastewater disposal is huge due to the large
volume of wastewater generated.
DIPA is highly water soluble. Thus, the removal of this
organic pollutant is tricky and very little literature is
available on this topic. Extracting the pollutant may
be one of the possible ways to solve the problem, but
the production of secondary waste may still be an
issue. In the past several years, advanced oxidation
processes (AOPs) have attracted many researchers.
Numerous reports on AOPs are available with a
handful of organic pollutants that are degradable
through this technique. According to Huang et al.
[1], Fenton’s reagent has been discovered over 100
years ago but its application as a destroying agent for
organic pollutants was only explored several decades
later. Among the applications of Fenton’s reagent that
have been reported included the degradation of azo
dye Amido black [2], textile effluents [3], cork cooking
[4], and pharmaceutical waste [5]. The Fenton system
is an attractive oxidant for wastewater treatment due
This paper was presented at the UK – Malaysian Engineering Conference 2008, London
14 -16 July 2008.
VOLUME Six NUMBER two july - december 2008 PLATFORM
21
Mission-Oriented Research: pETROCHEMICAL CATALYSIS TECHNOLOGY
to the fact that iron is a highly abundant and non-toxic
element, as well as the simple handling procedure and
environmentally benign of H2O2 [6].
Theory
Glaze et al. [7] defined AOPs as “near ambient
temperature and pressure water treatment processes
which involve the generation of hydroxyl radicals in
sufficient quantity to effect water purification”. The
main feature of AOP is the hydroxyl radical, ·OH, which
has a high oxidation potential and acts rapidly with
most organic compounds to oxidize them into CO2
and water. Hydroxyl radicals have the second largest
standard redox potential after fluorine, which is 2.8 V
[8].
H2O2. The direct photolysis of H2O2 leads to the
formation of the ·OH radical.
H2O2 →uv 2·OH
(4)
Combination of UV irradiation with Fenton’s system
known as photo-Fenton has also been a promising
technique in wastewater treatment and research has
been conducted on the application of this technique
to some organic compound. Based on the literature,
the presence of light has increased the production
rate of ·OH radical by an additional reaction as in (5).
Fe(OH)2+ →uv Fe2+ + ·OH
(5)
According to Matthew [8], although reactions (4) and
(5) are important, the most vital aspect of photoFenton is the photochemical cycling of Fe3+ back to
Fe2+.
Hydrogen peroxide (H2O2) is a strong oxidant, but
it alone is not effective in high concentration of
refractory contaminant such as amines at a reasonable
amount of H2O2. Thus, a relatively non-toxic catalyst
like iron was introduced to increase the rate of ·OH
radical production. Fenton’s reagent consists of
ferrous ion (Fe2+) and H2O2 which generates the ·OH
radical according to
There are few main factors affecting the process
efficiency. The concentration of H2O2, concentration
of FeSO4, UV power dosage, temperature and pH are
among the main contributing factors towards process
efficiency.
Fe2+ + H2O2 → Fe3+ + ·OH + OH−
Effect of H2O2 concentration
(1)
·OH radical may also be scavenged by reaction with
another Fe2+.
·OH + Fe2+ → Fe3+ + OH− (2)
The reaction between Fe3+ and H2O2 slowly
regenerated Fe2+.
In their report, Walling and Goosen [12] have simplified
the overall Fenton chemistry by considering the
dissociation of water as in Equation (3) which suggests
that the acidic environment is needed in order to
dissociate H2O2 by the presence of H+.
2Fe2+ + H2O2 + 2H+ → Fe3+ + 2 H2O
(3)
Another way to initiate the generation of ·OH radical
is by supplying UV radiation to the system containing
22
H2O2 plays an important role as the oxidising agent in
the AOP. Thus, it is vital to optimise the concentration
of H2O2 because the main cost of these methods is
the cost of H2O2 and excessive dosed of H2O2 trigger
side-effects [9] due to self-scavenging of ·OH radical
by H2O2 itself (Eqn. 6). Jian et al., [2] reported in their
study that the degradation of Amido black 10B was
reducing with high concentration of H2O2.
·OH + H2O2 → H2O + HO2·
(6)
Effect of the initial Fe2+ concentration
Optimum concentration of Fe2+ is vital for the reaction
as too low a concentration will slow the generation of
·OH radical thus reducing the removal of COD. But too
high a concentration could lead to self-scavenging of
·OH radical by Fe2+ (Eqn. 2) [2].
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Mission-Oriented Research: pETROCHEMICAL CATALYSIS TECHNOLOGY
Effect of UV power dosage
Hung at el. [9] in their report, wrote that COD removal
could be increased by increasing UV power dosage.
This is due to the faster formation rate of ·OH radical.
The dosage of UV could be controlled by the number
of lamps inside the reactor.
Effect of Temperature
Temperature is a critical factor to reaction rate,
production yield and distribution. Reaction rate is
expected to increase with increased temperature.
However, in these processes, no optimal temperature
was detected [10, 11]. Some researchers, however,
gave the opposite result. Anabela et al. [4] reported
in their literature that the optimal temperature was
30 °C in the degradation of cork cooking wastewater.
Effect of pH
Previous studies have shown that the acidic level of
near pH3 was usually the optimum. High values of pH
(>4) decreased the generation of ·OH radical because
of the formation of ferric hydroxo complexes but at
too low a pH value (<2), the reaction slowed down due
to the formation of [Fe(H2O2)6]2+ [11].
Research Methodology
Materials
DIPA was obtained from Malaysia Liquified Natural
Gas (MLNG); H2O2 and NaOH were from Systerm;
FeSO4·7H2O was from Hamburg Chemicals; and H2SO4
was from Mallinckrodt.
Experimental procedure
A stirred jacketed glass reactor was used to monitor
the progress of the reaction. The simulated waste of
DIPA in the desired concentration was prepared and
concentrated H2SO4 and 1M NaOH were added to
adjust the solution to the desired pH value. The ferrous
sulfate catalyst was added into the amine solution at
the beginning of the experiment and stirred to get a
homogeneous solution. This is then followed by the
addition of H2O2. The reaction started immediately
and the temperature was maintained by circulating
cooling water through the jacket. Samples were taken
at regular intervals of time for COD analysis. COD was
then measured by a Hach 5000.
For photo-Fenton oxidation, the experimental
procedure was similar to the Fenton process, except for
the additional UV irradiation. A quartz tube containing
a UV lamp 4W was inserted into the reactor.
Analysis
Samples of 3 ml in volume were taken and put into the
test tube containing 4 ml of 1M NaOH at the regular
interval for COD analysis. NaOH was added into the
samples to increase pH to 12 so that hydrogen peroxide
became unstable thus decomposing into oxygen
and water. Besides, this method can precipitate iron
into ferric hydroxide [Fe(OH)3] [12]. The precipitate
Fe(OH)3 was then separated from the solution by
using microfilter. In order to further ensure that no
interference of H2O2 to the COD measurement, the test
tubes containing samples were heated in the boiling
water for 10 minutes to remove the remaining H2O2
[13-15] as the peroxide is also unstable at temperature
higher than 40 °C. The level of reproducibility for this
system is high where the COD percentage removal
will only vary about 5 percent between runs for the
same parameters.
Result and Discussion
Effect of H2O2 Concentration
The effect of H2O2 concentration on COD removals
was examined by changing the H2O2 concentration
while keeping the concentration of FeSO4, pH and
temperature constant at 4.98 g/L, 3 and 30 °C,
respectively (Fig. 1).
Figure 1(a) shows the percentage of COD removal
versus time for different H2O2 concentration at initial
COD of 50,000 mg/L. From the figure, the COD
removal increases with increasing H2O2 concentration.
VOLUME Six NUMBER two july - december 2008 PLATFORM
23
Mission-Oriented Research: pETROCHEMICAL CATALYSIS TECHNOLOGY
Fe:H2O2 = 1:20
30
Fe:H2O2 = 1:30
Fe:H2O2 = 1:40
50
28
Fe:H2O2 = 1:50
40
COD Removal (%)
COD Removal (mg/L)
Fe:H2O2 = 1:60
30
20
26
24
22
20
10
18
0
1:20
0
10
20
30
40
50
Time (min)
60
70
80
1:30
90
(a)
1:40
1:50
1:60
Fe: H2O2
(b)
Figure 1. Effect of H2O2 concentration on COD removal for Fenton Oxidation (Initial COD = 50,000 mg/L; FeSO4 was 4.98 g/L; temperature
= 30 °C, pH = 3)
However, when the Fe:H2O2 ratio is more than 1:50,
the percentage removal decreases as in Figure 1(b).
This may be due to the scavenging effect of H2O2 as
in Eqn. (6). When too much H2O2 was in the solution,
it reacted with the ·OH radical and subsequently
reduced the concentration of ·OH radical available to
attack the organic compound.
Figure 2(a) shows the percentage of COD removal by
using photo-Fenton oxidation method. It followed
the same trend as Fenton’s system. The percentage of
COD removal increased when the hydrogen peroxide
concentration increased. But when the Fe:H2O2
ratio was more than 1:30, the percentage removal
decreased as in Figure 2(b).
·OH + H2O2 → H2O + HO2·
The efficiency between Fenton’s system and photoFenton was compared as in Figure 3. From the figure,
(6)
35
26
25
25
COD Removal (%)
COD Removal (%)
30
20
15
Fe: H2O2 = 1:20
Fe: H2O2 = 1:30
10
Fe: H2O2 = 1:40
0
Fe: H2O2 = 1:60
0
10
20
30
40
50
60
70
80
22
21
19
1:20
90
Time (min)
(a)
23
20
Fe: H2O2 = 1:50
5
24
1:30
1:40
1:50
1:60
Fe:H 2O2
(b)
Figure 2. Effect of H2O2 concentration on COD removal for photo-Fenton Oxidation (Initial COD = 50,000 mg/L; FeSO4 was 4.98 g/L;
temperature = 30 °C, pH = 3)
24
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Mission-Oriented Research: pETROCHEMICAL CATALYSIS TECHNOLOGY
for both Fenton and photo-Fenton was 1:50. COD
removal of 31% and 24% was achieved for Fenton
and photo-Fenton, respectively. Initial comparison
between Fenton and photo-Fenton showed that
photo-Fenton gave better degradation. However,
this is not conclusive and further research will have
to be conducted to increase the percentage of COD
removal and reduce sludge formation.
References
Figure 3. Comparison of degradation efficiency between Fenton
and photo-Fenton oxidation of DIPA (Initial COD=50,000 mg/L;
FeSO4 was 4.98 g/L; temperature = 28 °C, pH = 3)
the percentage COD removal for photo-Fenton is
slightly higher compared to Fenton. This could be due
to the additional UV irradiation added into the system.
The addition led to the additional production of ·OH
radical which then increased the concentration of ·OH
radical. Besides, UV irradiation was able to regenerate
ferrous catalyst by the reduction of Fe(III) to Fe(II) as in
Eqn. (7).
Fe(III)OH2+ →uv Fe(II) + ·OH
[1]
C. P. Huang, C. Dong, Z. Tang, “Advanced chemical oxidation:
its present role and potential future in hazardous waste
treatment”, Waste Mgmt, 13 (1993) 361-577
[2]
J. H. Sun, S. P. Sun, G. L. Wang, and L. P. Qiao (2006)
“Degradation of azo dye Amido black 10B in aqueous
solution by Fenton oxidation process”, Dyes and Pigment,
B136, 258-265G
[3]
M. Perez, F. Torrades, X. Domenech, J. Peral (2001) “Fenton
and photo-Fenton oxidation of textile effluents”, Water
Research 36, 2703-2710
[4]
A. M. F. M. Guedes, L. M. P. Madeira, R. A. R. Boaventura and
C. A. V. Costa (2003) “ Fenton oxidation of cork cooking
wastewater-overall kinetic analysis” , Water Research 37,
3061-3069
[5]
Huseyin, T., Okan, B., Selale, S. A., Tolga, H. B., I. Haluk
Ceribasi, F. Dilek Sanin, Filiz, B. D. and Ulku, Y. (2006) “ Use
of Fenton oxidation to improve the biodegradability of a
pharmaceutical wastewater” , Hazard. Mater. B136 (258265)
[6]
Rein, M (2001) “Advanced oxidation processes – current
status and prospects” , Proc. Estonian Acad. Sci. Chem., 50,
2, 59-80
[7]
Glaze, W. H., Kang, J. W. and Chapin, D. H. (1987) “ The
chemistry of water treatment processes involving ozone,
hydrogen peroxide and ultraviolet radiation”, Ozone Science
Engineering
[8]
Matthew, A. T., (2003) “ Chemical degradation methods
for wastes and pollutants, environmental and industrial
applications” , pp. 164-194, Marcel Dekker Inc
[9]
Hung, Y. S, Ming, C. C. and Wen, P. H. (2006) “ Remedy of
dye manufacturing process effluent by UV/H2O2 process”,
Hazard. Mater. B128, 60–66
(7)
However, at ratio Fe:H2O2 of 1:50, the result obtained
was different. The percentage COD removal for Fenton
was higher. The problem could be due to the stirring
method. The stirrer used was small and the volume
of the solution inside the reactor was high. Besides,
the presence of a quartz tube inside the reactor might
have reduced the stirring efficiency. One way to
overcome the problem is by using aeration to mix the
solution and this is recommended for further research
plans.
Conclusion
The applicability of Fenton and photo-Fenton for the
degradation of diisopropanolamine was investigated.
By keeping the initial COD, concentration of FeSO4, pH
and temperature constant at 50,000 mg/L, 4.98 g/L, 3
and 30 °C respectively, the optimum ratio of Fe:H2O2
[10] Dutta, K., Subrata, M., Sekhar B., and Basab C. (2001) @
Chemical oxidation of methylene blue using a Fenton-like
reaction@ , Hazard Mater. B84, 57-71
VOLUME Six NUMBER two july - december 2008 PLATFORM
25
Mission-Oriented Research: pETROCHEMICAL CATALYSIS TECHNOLOGY
[11] Ipek, G., Gulerman, A. S. and Filiz, B. D. (2006) “Importance
of H2O2/Fe2+ ratio in Fenton’s treatment of a carpet dyeing
wastewater”, Hazard. Mater. B136, 763-769
[12] C. Walling and A. Goosen, (1973) “Mechanism of the ferric
ion catalysed decomposition of hydrogen peroxide: effects
of organic substrate”, J. Am. Chem. Soc. 95 (9) 2987-2991
[13] Kavitha, V. and Palanivelu, K. (2005) “The role of ferrous ion
in Fenton and photo-Fenton processes for the degradation
of phenol”, Chemosphere 55, 1235-1243
[14] Lou J. C. and Lee S. S. (1995) “Chemical Oxidation of BTX
using Fenton’s reagent”, Hazard Mater. 12, No 2, 185-193
[15] Jones C. W. (1999) “Introduction to the preparation and
properties of hydrogen peroxide”. In: Clark, J. H. (Ed)
Application of Hydrogen Peroxide and Derivatives. Royal
Society of Chemistry, Cambridge, UK, pp. 30
Abdul Aziz Omar is currently Head of the
Geosciences and Petroleum Engineering
Department,
Universiti
Teknologi
PETRONAS.
Associate Professor Aziz completed his
tertiary education in the United States,
where he obtained his Master’s and
Bachelor’s degrees from Ohio University
in 1982. While his undergraduate training was in Chemistry
and Chemical Engineering, his graduate specialisation was in
Environmental Studies. He has over 15 years of experience as an
academician, 6 years as a process/project engineer and 4 years as
a senior manager. He has also worked on many projects related
to EIA (Environmental Impact Assessment), safety studies and
process engineering design. Among his many experiences, one
significant one would be the setting up of the School of Chemical
Engineering at Universiti Sains Malaysia. He was appointed the
founding Dean of the School. In March 2001, he joined Universiti
Teknologi PETRONAS (UTP) as a lecturer, where he continues
to teach and pursue his research interests. Assoc. Prof. Aziz is a
Professional Engineer in Chemical Engineering, registered with
the Malaysian Board of Engineers since 1989, and a Chartered
Engineer in United Kingdom from 2006. He is a Fellow of the
Institution of Chemical Engineers (IChemE), UK and is a Board
member of IChemE in Malaysia.
26
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Mission-Oriented Research: pETROCHEMICAL CATALYSIS TECHNOLOGY
Synthesis of well-defined iron nanoparticles
on a spherical model support
Noor Asmawati Mohd Zabidi*, P. Moodley1, P. C. Thüne1, J. W. Niemantsverdriet1
*Universiti Teknologi PETRONAS, 31750 Tronoh, Perak Darul Ridzuan, Malaysia
1Schuit Institute of Catalysis, Eindhoven University of Technology
*noorasmawati_mzabidi @petronas.com.my
Abstract
Spherical model catalysts consisting of SiO2-sphere-supported iron nanoparticles were prepared using the
colloidal synthesis approach, the reverse microemulsion and the ammonia deposition methods. Amongst these
synthesis methods, the colloidal synthesis approach was found to be the most promising synthesis route for the
Fe/SiO2 model catalysts. The modified colloidal synthesis method produced nearly monodisperse sphericalshaped iron oxide nanoparticles with average diameters of 6.2 ± 0.9 nm on the SiO2 spheres. X-ray photoelectron
spectroscopy (XPS) analyses revealed that the catalyst contained Fe2O3 (hematite). Morphological changes were
observed on the spherical Fe/SiO2 model catalysts during the Fischer-Tropsch synthesis (FTS) reaction.
Keywords. Nanoparticles, iron, spherical model catalyst, Fischer-Tropsch reaction
Introduction
Iron has been the traditional catalyst of choice for
the Fischer-Tropsch synthesis due to its favorable
economics. However, knowledge on the relation
between the rate of the reaction to the composition
and morphology of the catalyst is still lacking [1].
The use of spherical model catalyst system enables
investigation on the fundamental aspects of the
catalyst, such as influence of particle size, phase and
composition on the catalytic activity [2]. The objective
of the present work is to prepare and characterize
spherical model SiO2-supported iron catalysts. The
catalysts were prepared using the colloidal synthesis
approach [3], the reverse microemulsion method
[4] and the ammonia deposition method [5]. The
colloidal synthesis approach was adapted from a
method described by Sun and Zeng [3] which involved
homogeneous nucleation process. However, our aim is
to stabilise the iron nanoparticles on the SiO2 spheres
through a heterogeneous nucleation process.
The usage of spherical model silica support allows
for viewing of iron nano-particles in profile with
transmission electron microscopy. Supported
iron nano-particles in combination with electron
microscopy are well suited to study morphological
changes that occur during the Fischer-Tropsch
synthesis. The spherical model catalyst enables
investigation on the fundamental aspects of the
catalyst, such as influence of particle size, phase and
composition on the catalytic activities. This paper
presents the results of three synthesis approaches for
the spherical Fe/SiO2 model catalysts as well as their
morphological changes upon exposure to syngas.
This paper was presented at the International Conference on Nanoscience and Nanotechnology 2008, Shah Alam,
18 - 21 November 2008
VOLUME Six NUMBER two july - december 2008 PLATFORM
27
Mission-Oriented Research: pETROCHEMICAL CATALYSIS TECHNOLOGY
Experimental
For the colloidal synthesis method [3], non-porous
silica spheres (BET surface area = 23 m2g-1, pore
volume = 0.1 cm3g-1 and diameter = 100 - 150 nm)
were sonicated in a mixture of olelylamine, oleic
acid and cyclohexane for 1 hour and then heated
and stirred in a multi-neck quartz reaction vessel. A
liquid mixture of iron(III) acetyl acetonate, oleylamine,
oleic acid, 1,2-hexadecanediol, and phenyl ether was
slowly added to the stirred SiO2 suspension once the
reaction temperature reached 150 °C. The reaction
mixture was refluxed under nitrogen atmosphere at
265 °C for 30 minutes.
The reverse microemulsion method [4] involved
preparing two reverse microemulsions. The first
reverse microemulsion consisted of Fe(NO3)3.9H2O
(aq) and sodium bis-(2-ethylhexyl) sulfosuccinate
(AOT, ionic surfactant) in hexanol. The second reverse
microemulsion was prepared by mixing an aqueous
hydrazine solution (reducing agent) with the AOT
solution. SiO2 spheres were added to the mixture
and the slurry was stirred for 3 hours under nitrogen
environment. The ammonia deposition method
utilised Fe(NO3)3.9H2O and 25 wt% aqueous ammonia
[5].
The calcined catalyst samples were placed on carboncoated Cu grids for characterisation by transmission
electron microscopy. TEM studies were carried
Figure 1(a)
out on a Tecnai 20 (FEI Co) transmission electron
microscope operated at 200 kV. XPS was measured
with a Kratos AXIS Ultra spectrometer, equipped with
a monochromatic Al Kα X-ray source and a delay-line
detector (DLD). Spectra were obtained using the
aluminium anode (Al Kα = 1486.6 eV) operating at
150 W. Spectra were recorded at background pressure,
2 x 10 -9 mbar. Binding energies were calibrated to Si2s
peak at 154.25 eV.
The activities of the spherical model catalysts for
Fischer-Tropsch synthesis were evaluated in a fixedbed quartz tubular microreactor equipped with an
on-line mass spectrometer (Balzers). Catalyst samples
were pre-reduced in H2 at 450 °C for 2 h and then
exposed to syngas (H2:CO = 5:1) at 270 °C. The samples
were regenerated via heating in a flow of 20% oxygen
in helium up to 600 °C.
Results and discussion
Figures 1(a), (b) and (c) show the TEM images of
spherical model catalysts comprising iron oxide nanoparticles anchored on SiO2 spheres prepared via the
modified colloidal synthesis approach, the reverse
microemulsion method and the ammonia deposition
method, respectively. Spherical-shaped iron oxide
nano-particles with average diameters of 6.2 ± 0.9
nm were formed via the modified colloidal synthesis
method and the nano-particles were almost evenly
dispersed on the SiO2 surfaces. An equimolar mixture
Figure 1(b)
Figure 1(c)
Figure 1. TEM images of calcined catalysts of iron oxide nanoparticles on SiO2 spheres prepared via the (a) colloidal synthesis approach
(b) reverse microemulsion method (c) ammonia deposition method.
28
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Mission-Oriented Research: pETROCHEMICAL CATALYSIS TECHNOLOGY
Intensity (AU)
(c)
(b)
(a)
275
280
285
290
295
300
Binding energy (eV)
Figure 2. TEM image of iron oxide nanoparticles on
SiO2 spheres after the Fischer Tropsch reaction at 270
°C for 2 hours at H2:CO ratio of 5:1.
Figure 3. XPS showing carbon region for (a) fresh (b) regenerated (c) spent
Fe/SiO2
of oleylamine and oleic acid was used in the colloidal
synthesis approach and these surfactants were able
to prevent the agglomeration of the iron oxide nanoparticles. Iron oxide nano-particles were anchored
on the SiO2 surfaces and did not lie in between the
SiO2 spheres, as shown in Figure 1(a), thus suggesting
that nucleation occurred heterogeneously. The iron
loading was kept at 6 wt% as we have discovered
that increasing the iron loading resulted in highly
agglomerated nano-particles. The size of the nanoparticle is influenced by temperature, time, surfactant,
amounts of metal precursor as well as the ratio of the
metal precursor to the surfactant [6]. The reverse
microemulsion method also produced sphericalshaped iron oxide nano-particles with average
diameters of 6.3 ± 1.7 nm, however, the coverage
of the SiO2 surfaces was found to be less than that
obtained using the colloidal synthesis approach. The
result of the synthesis via the ammonia deposition
method showed extensive agglomeration of the iron
nanoparticles, as depicted in Figure 1(c).
Table 1. Atomic ratios based on XPS analyses
Sample
The spherical model catalysts synthesised by
the colloidal synthesis method and the reverse
microemulsion method were tested in a Fischer Tropsch
reaction. However, only the catalyst synthesised via
the colloidal method showed some activities in the
Fischer Tropsch reaction. Changes on the morphology
were investigated upon exposure to the syngas. Figure
2 shows the TEM image of the catalyst after two hours
exposure to the syngas. It shows ~ 50% increase in the
size of the nano-particles and formation of an outer
rim of thickness 3.2 ± 0.6 nm, following exposure to
the syngas. A darker color at the centre of the used
catalyst nano-particles suggests that the iron oxide
remained in the core whereas the outer rim consists
of amorphous carbon (EB = 284.5 eV), as confirmed by
the XPS analyses (Figure 3). Table 1 shows the atomic
ratios obtained from XPS analyses. Figure 4 shows the
presence of Fe3p peak at EB= 56.0 eV for the fresh and
the used catalyst, thus suggesting that the catalyst
remained as Fe2O3 (hematite). Upon contact with H2/
CO, oxygen-deficient Fe2O3 was observed, however,
the oxygen vacancies did not reach a critical value
that can lead to nucleation of Fe3O4.
O1s Fe / O1s Si
Fe 3p / O Fe2O3
C1s / Fe3p
Fresh
0.248
0.186
2.02
Spent
0.199
0.066
7.64
Regenerated
0.231
0.076
6.24
VOLUME Six NUMBER two july - december 2008 PLATFORM
29
Mission-Oriented Research: pETROCHEMICAL CATALYSIS TECHNOLOGY
Fe 3p
Acknowledgements
The authors thank Mr. Tiny Verhoeven for the TEM and XPS
measurements. The authors would also like to thank Mr. Denzil
Moodley for his assistance in the activity studies. We acknowledge
the financial support for this project from Sasol South Africa.
Noor Asmawati Mohd Zabidi acknowledges the support given
by Universiti Teknologi PETRONAS under the sabbatical leave
programme.
Fe 3p
Fresh
Regenerated
Spent
65
Fresh
70
AU
60
ergy (eV)
Regenerated
References
Spent
45
50
55
60
65
70
[1]
A. Sarkar, D. Seth, A.K. Dozier, J.K. Neathery, H. H. Hamdeh,
and B. H. Davis, “Fischer-Tropsch synthesis: morphology,
phase transformation and particle size growth of nanoscale particles”, Catal. Lett. 117 (2007) 1
[2]
A. M. Saib, A. Borgna, J. van de Loosdrecht, P. J. van Berge,
J. W. Geus, and J. W. Niemantsverdriet, “Preparation and
characterization of spherical Co/SiO2 model catalysts
with well-defined nano-sized cobalt crystallites and a
comparison of their stability against oxidation with water”,
J. Catal. 239 (2006) 326
[3]
S. Sun and H. Zeng, “Size-controlled synthesis of magnetite
nanoparticles”, J. Am. Chem. Soc. 124 (2002) 8204
[4]
A. Martinez and G. Prieto, “The key role of support surface
tuning during the preparation of catalysts from reverse
micellar-synthesized metal nanoparticles”,Catal. Comm. 8
(2007) 1479
[5]
A. Barbier, A. Hanif, J.A. Dalmon, G.A. Martin, “Preparation
and characterization of well-dispersed and stable Co/SiO2
catalysts using the ammonia method”, Appl. Catal. A. 168
(1998) 333
[6]
T. Hyeon, “Chemical synthesis of magnetic nanoparticles,
”Chem. Commun. (2003) 927
Binding energy (eV)
Figure 4. XPS showing Fe3p region for (a) fresh (b) spent (c)
regenerated Fe/SiO2
Conclusions
Spherical model catalysts consisting of SiO2supported iron nano-particles have been prepared
and characterized using TEM and XPS. The modified
colloidal synthesis approach resulted in sphericalshaped iron oxide nano-particles with average
diameters of 6.2 ± 0.9 nm. The modified colloidal
synthesis method produced better dispersion of the
iron oxide nano-particles compared to that obtained
from the reverse microemulsion method. The spherical
Fe/SiO2 model catalyst prepared via the modified
colloidal synthesis method exhibited some activity
towards Fischer-Tropsch synthesis whereas the one
synthesized via the reverse microemulsion method
showed negligible activity for FTS. Morphological
changes were observed on the spherical Fe/SiO2 model
catalysts upon exposure to syngas and during the reoxidation step. TEM results show ~ 50% increase in the
size of the iron oxide nano-particles and formation of
the carbon rim around the iron oxide nano-particles
(confirmed by XPS) upon a 2-hour exposure to the
syngas. XPS measurements confirmed the presence of
Fe2O3 nano-particles in the fresh and the used catalyst
samples. The results of our investigation show that
the spherical Fe/SiO2 model nano-catalyst of welldefined size can be prepared and characterised. This
can facilitate the size-dependent studies of the ironbased catalyst in FTS.
30
Noor Asmawati Mohd Zabidi obtained
her PhD in 1995 from University of MissouriKansas City, USA. She joined Universiti
Teknologi PETRONAS in 2001 and was
promoted to an Associate Professor in
2005. She was on a sabbatical leave at
Eindhoven University of Technology, The
Netherlands from March–December 2007.
During the sabbatical leave, she carried out
a research project on the synthesis of SiO2-supported iron
nanocatalysts for the Fischer Tropsch reaction. Her research
interests are photocatalysis and catalytic conversion of gas to
liquid (GTL) fuels and chemicals.
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: Fuel combustion
Performance and Emission Comparison of A DirectInjection (DI) Internal Combustion Engine using
Hydrogen and Compressed Natural Gas as Fuels
A. Rashid A. Aziz*, M. Adlan A., M. Faisal A. Muthalib
Universiti Teknologi PETRONAS, 31750 Tronoh, Perak Darul Ridzuan, Malaysia
*[email protected]
Abstract
Hydrogen internal combustion engine is considered as a suitable pathway to hydrogen economy before fuel cell
technologies became more mature and cost effective. In this study, combustion of hydrogen and compressed
natural gas (CNG) in a direct-injection single cylinder research engine was investigated. Engine performance
parameters such as the power, torque, BMEO and COV of hydrogen DI operation were measured and compared
to CNG-DI operation. Stoichiometric combustion of CNG at part load (50% throttle opening) is presented and
compared with hydrogen at 0.2 and 0.8 equivalent ratio. The slightly-lean hydrogen (0.8 equivalent ratio) resulted
in a better overall performance and emission of the engine.
introduction
Global warming and air pollution issues have
brought international efforts to scale down the use
of hydrocarbon fuel, which is one of the biggest
contributors to a number of greenhouse gases.
On the other hand, fossil fuel reserves, especially
petroleum, is depleting and will be at its peak in just
a couple of decades while the demand from industry
and transportation is increasing [1].
Hydrogen has been introduced as an alternative energy
carrier. It can be made from both renewable and fossil
energy. By using renewable energy or nuclear power
plant to produce hydrogen, greenhouse gases can
be totally eliminated. It is better than electricity in
terms of distribution efficiency, refuelling speed and
energy capacity. On the other hand, hydrogen vehicle
performance is comparable to hydrocarbon-fuelled
vehicles. In addition, the only emission of hydrogen
is water vapour.
However, the current hydrogen engines still face
practical problems that mask the actual capability
of hydrogen as fuel for transportation. Hydrogen
engine’s power and speed are limited due to knock,
backfire [2], low volumetric efficiency [3] and a number
of other problems. Some researchers suggested
the use of a direct-ignition system and a number of
experiments have been conducted that provided
proof that using hydrogen as fuel produces better
output than its gasoline counterparts [2].
In this study, the main objective was to examine
the performance of a hydrogen-fuelled directinjection spark ignition internal combustion engine.
Aspects that were observed were power and torque
output, emissions, fuel consumption and operating
characteristics.
Experiments
Experiments were done using a four-stroke Hydra
Single Cylinder Optical Research Engine at the
This paper was presented at the International Gas Research Union Conference 2008, Paris
8 - 10 October 2008
VOLUME Six NUMBER two july - december 2008 PLATFORM
31
Technology Platform: fuel combustion
Centre for Automotive Research, Universiti Teknologi
PETRONAS, Malaysia. The specification of the engine
is listed in Table 1. Engine control parameters such
as injection timing, ignition timing and air-fuel ratio
were controlled by an ECU that was connected to a
computer. All output parameters of the engine was
obtained from a high-speed data acquisition as well
as a low-speed data acquisition from the engine
dynamometer control interface.
Two experiments were conducted in this study –
“ultra-lean combustion” (equivalence ratio of 0.2)
and “slightly-lean combustion” (0.8 which is near
stoichiometric). The first experiment used lowpressure injector (7.5 bar). However, because the
flow-rate of that injector was too slow, a high-pressure
injector had to be used for the second experiment.
Both experiments used the stratification method,
also known as stratified-lean combustion. Figure 1
illustrates a piston with 35 mm bowl that was used
in the experiments and Figure 2 shows the position
of the injector relative to the spark plug in the direct
injection system. Table 2 lists the main operating
parameters.
Table 1. Engine details
Besides comparing results from both experiments,
another result from an experiment on natural gas fuel
was also made (Table 3).
Figure 1. Stratified-lean piston
Figure 2. Location of injector
and spark-plug relative to the
piston at TDC
Table 2. Main parameters of the two experiments
Engine type
4-stroke spark ignition
No. of cylinders
One
Parameter
Ultra-lean
Slightly-lean
Displacement volume
399.25 cm3
Equivalence ratio
0.2
0.8
Cylinder bore
76 mm
Injector Rail
Pressure
7.5 bar
18 bar
Cylinder stroke
88 mm
Injection Timing
130 deg BTDC
130 deg BTDC
Compression ratio
14:1
Load
Part throttle
Part throttle
Exhaust valve open
10° ATDC
Ignition Timing
MBT
MBT
Exhaust valve closed
45° BBDC
Inlet valve open
12° BTDC
Inlet valve closed
48° ABDC
Fuel induction
Direct-injection
Rail Pressure
Value
Table 3. Main parameters of the CNG-DI experiment
Parameter
Value
Equivalence ratio
Stoichiometric
Injector Rail Pressure
18 bar
7.5 and 18 bar
Injection Timing
130 deg BTDC
Injector position
Centre
Load
Part throttle
Injector nozzle type
Wide angle
Ignition Timing
MBT
32
PLATFORM VOLUME Six NUMBER TWO july - december 2008
50
40
2500 rpm
3000 rpm
35
14
30
0.8 Ign.
15
6
10
4500 rpm
0
-5
20
30
40
Crank Angle (deg ATDC)
5000
0
4500
5
2
4000
4
3500 rpm
4000 rpm
3500
10
20
0.2 Ign.
8
3000
0
10
2500
-10
0.8 CoV
2000
20
25
0.2 CoV
1500
30
10
-20
16
12
CoV (%)
Cylinder Pressure (bar)
60
Ignition (deg BTDC)
Technology Platform: Fuel combustion
Engine Speed (rpm)
Figure 3. Pressure developed during compression and expansion
stroke at various engine speeds for 0.8 equivalence ratio
Figure 4. Coefficient of variation at different engine speed with
its corresponding spark advance
results and discussions
A study showed that MBT ignition advance increases
when equivalence ratio decreases [5]. For a mixture of
0.2 equivalence ratio, the MBT ignition timing range
is between 30 to 50 degrees BTDC. This is consistent
with the current results.
The engine was unstable in ultra-lean mode. It
alternated between producing positive torque and
negative torque – showing that the engine was not
producing positive work. Peak pressure varied as
much as 13 percent (Figure 3).
Possible cause of the variation could be from misfiring.
The injector that was used for this experiment was a
low pressure injector with low mass flow rate – about
2.7 ms to fully inject the fuel. At 3 500 rpm, this
corresponded to 56.7 CA degree which was a long
duration for hydrogen injection. A study showed that,
for equivalence ratio between 0.7 to 1.4, the coefficient
of variation (CoV) is low but increases significantly
when the ratio moves far from the range. The study
also concluded that cycle variation is caused by
variation in the early combustion period [4]. Running
the engine at slightly-lean resulted in lower CoV.
The combustion produced very high peak pressure,
common for hydrogen engine because of its high
flame temperature and high flame speed. The peak
pressure reached up to 60 bars (Figure 4) but lower
than what actually the engine could produce. It was
seen that the ignition timing was set near TDC (Figure
3) to achieve MBT. Advancing the ignition resulted in
a peak pressure reaching up to 90 bars with knocks
occurring in the cylinder.
Peak pressures of natural gas combustion nearly
doubled the peak pressure for slightly-lean hydrogen
combustion (Figure 5). Despite the high peak pressure,
natural gas was still running without knock as opposed
to hydrogen.
Performance
Combustion of ultra-lean mixture faced consistent
misfiring which resulted in very low BMEP (Figure 6).
Judging from the CoV, when the speed was increased,
8
6
BMEP (bar)
Engine Stability
Ultra-lean H2
Slightly-lean H2
4
CNG
2
0
1500
2000
2500
3000
3500
4000
4500
5000
Engine Speed (rpm)
Figure 5. BMEP comparison of ultra-lean H2, slightly-lean H2 and
stoichiometric natural gas
VOLUME Six NUMBER two july - december 2008 PLATFORM
33
Technology Platform: fuel combustion
misfiring occurred frequently (Figure 3). BMEP
decreased until it could no longer produce any work
after 4 500 rpm.
Comparison at 3500 rpm
60
Pressure (bar)
50
H2 - 0.8
CNG
H2 - 0.2
40
30
20
10
0
-10
0
0.0001
0.0002
0.0003
0.0004
0.0005
Usage of a wide angle injector caused only small
amounts of fuel to be concentrated inside the piston
bowl, which was less than ignitable air-fuel ratio. A
possible solution is to utilise a narrow angle injector
with stoichiometric-lean piston. A stoichiometric lean
piston has a smaller bowl and could concentrate the
mixture to be around the stoichiometric ratio. [6]
Volume (m3)
Figure 6. Pressure map comparison of ultra-lean H2, slightly-lean
H2 and stoichiometric natural gas at 3500 rpm
Brake-Torque (Nm)
30
25
20
15
Ultra-lean H2
10
Slightly-lean H2
CNG
5
0
1500 2000 2500 3000 3500 4000 4500 5000
Engine Speed (rpm)
For the slightly-lean mixture, the usage of late ignition
hampered its potential performance. At TDC, when the
pressure is around 30 to 40 bars, the pressure diagram
(Figure 4) clearly shows that rather than a smooth line,
the pressure line dropped and ended up with lower
peak pressure. In Figure 7, it was seen that near TDC,
the pressure drop slightly after ignition, which was
caused by the ignition delay.
In comparison with natural gas, hydrogen produced
more torque and power. For torque, as well as BMEP
curve, the values decreased with increase in engine
speed (Figures 6 & 8). However, the power output
of slightly-lean hydrogen did not show a drop with
increase speed (Figure 9).
Figure 7: Brake-torque comparison of ultra-lean H2, slightly-lean
H2 and stoichiometric natural gas
Comparison at 4500 rpm
10
Brake-Power (kW)
70
Pressure (bar)
60
50
H2 - 0.8
40
H2 - 0.2
CNG
30
20
8
6
4
2
Ultra-lean H2
Slightly-lean H2
CNG
10
0
0
-10
1500 2000 2500 3000 3500 4000 4500 5000
0
0.0001
0.0002
0.0003
0.0004
0.0005
Engine Speed (rpm)
Volume (m 3)
Figure 8: P-V diagram of a cycle at 4500 rpm and 0.8 equivalence
ratio.
34
Figure 9. Brake-power comparison between ultra-lean and
slightly-lean hydrogen with stoichiometric natural gas.
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: Fuel combustion
Fuel Economy
1.4
The engine’s fuel consumption at ultra-lean mixture
is fairly constant before 4 000 rpm (Figure 10). The
sharp increase after that point could be linked to the
misfiring that became worse at higher speeds. As the
speed increases, the fuel had less time to mix with
the air, making a heterogeneous mixture. At slightlylean mixture, the consumption of fuel was also low
– primarily caused by a stable combustion without
misfire.
BSFC (kg/kW-h)
1.2
Ultra-lean H2
1.0
Slightly-lean
H2
0.8
0.6
0.4
0.2
0.0
1500 2000 2500 3000 3500 4000 4500 5000
Engine Speed (rpm)
Figure 10. Brake specific fuel consumption (BSFC) at various
engine speeds.
1.2
NOx (g/m3)
1.0
0.8
Figure 11 shows the fuel consumption comparison
between hydrogen with natural gas. The graph
shows the BSFC in equivalent gasoline consumption.
Conversion was based on the fuels’ lower heating
value. From the graph, it was seen that slightly-lean
hydrogen had lower BSFC than natural gas. The ultralean hydrogen performed poorly at higher engine
speeds.
0.6
0.4
Ultra-lean H2
Slightly-lean H2
0.2
CNG
Emissions
0.0
1500 2000 2500 3000 3500 4000 4500 5000
Engine Speed (rpm)
Figure 11. NOx comparison between ultra-lean and slightly-lean
hydrogen with stoichiometric natural gas.
Gasoline Equivalent BSFC
(kg/kW-h)
4.0
3.5
3.0
2.5
2.0
Ultra-lean H2
Slightly-lean H2
CNG
1.5
1.0
0.5
0.0
1500 2000 2500 3000 3500 4000 4500 5000
Engine Speed (rpm)
Figure 12. Comparison of BSFC in term of gasoline usage.
The only significant emission for hydrogen engine
is nitrogen oxides (NOx). However, the comparison
shows that natural gas emits more NOx than hydrogen
(Figure 12). At ultra-lean, the amount of emitted
NOx is considerably high especially at higher engine
speeds. The existence of NOx is usually related to
high combustion temperature in the cylinder. As
speed increases, there is less time for heat to transfer
from cylinder to the atmosphere, which increases
temperature (Figure 13).
The test-bed did not have an in-cylinder thermocouple
which could be used to determine temperatures inside
the cylinder. However, the system did have an exhaust
gas temperature. The exhaust temperature value was
used to show a rough estimation of the combustion
cylinder temperature.
Comparison in Figure 14 shows that the amount of
carbon dioxide emitted by hydrogen engine is really
insignificant compared to hydrocarbon fuel. The large
percentage of carbon dioxide in natural gas exhaust is
mainly caused by the carbon element in the structure
VOLUME Six NUMBER two july - december 2008 PLATFORM
35
Technology Platform: fuel combustion
conclusions
6
CO2 (vol %)
5
4
3
2
Ultra-lean H2
Slightly-lean H2
1
CNG
0
-1
1500
2000 2500 3000 3500 4000
Engine Speed (rpm)
4500 5000
Figure 13. CO2 comparison between ultra-lean and slightly-lean
hydrogen with stoichiometric natural gas.
120
100
950
NOx
Temp.
925
900
80
875
60
850
40
825
20
800
Temperature (K)
NOx (mg/m3)
140
0
775
1500 2000 2500 3000 3500 4000 4500 5000
Based on the above results and discussions, the
following conclusions were derived:
• Direct-injection could avoid the backfire
phenomenon and reduce the likelihood of preignition.
• At part load, power output of hydrogen was better
than natural gas. This proved that the actual
power output of hydrogen is higher than current
commercial fuels.
• Hydrogen at slightly-lean mixture has better fuel
economy than ultra-lean mixture.
• The only significant emission of hydrogen engine
is NOx but it is still lower than the amount emitted
by natural gas.
• The amount of CO2 by hydrogen engine is much
less than natural gas.
REFERENCES
Engine Speed (rpm)
[1]
Rifkin, J. (2002). “When There is No More Oil…: The
Hydrogen Economy”. Cambridge, UK: Polity Press
Figure 14. Temperature of exhaust gas and the amount of emitted
NOx at various speeds for slightly-lean combustion.
[2]
Das, L. M. (1990). “Fuel induction techniques for a hydrogen
operated engine”. International Journal of Hydrogen
Energy, 15, 833-842
[3]
Yi, H. S., Lee S. J., & Kim, E. S. (1996). “Performance evaluation
and emission characteristics of in-cylinder injection type
hydrogen fueled engine”. International Journal of Hydrogen
Energy, 21, 617-624
[4]
Kim, Y. Y., Lee J. T. & Choi, G. H. (2005). “An investigation on
the causes of cycle variation in direct injection hydrogen
fueled engines”. International Journal of Hydrogen Energy,
30, 69-76.
[5]
Mohammadi, A., Shioji, M., Nakai, Y., Ishikura, W. & Tabo, E.
(2007). “Performance and combustion characteristics of a
direct injection SI hydrogen engine”. International Journal
of Hydrogen Energy, 32, 296-304
[6]
Zhao, F. F., Harrington, D. L. & Lai, M. C. (2002). “Automotive
Gasoline Direct-Injection Engines”. Warrendale, PA: Society
of Automotive Engineers
[7]
Norbeck, J. M., et al. (1996). “Hydrogen Fuel for Surface
Transportation”. Warrendale, PA: Society of Automotive
Engineers
of methane – a major constituent of natural gas. On
the other hand, the existence of carbon dioxides in
hydrogen fuelled engine is mainly contributed by
combustion of engine oil on the cylinder wall [7].
The negative value of CO2 on the graph occurred
because the emitted CO2 was less in relative to the
calibration value. The system calculated CO2 in
percentage of total exhaust gas – not in ppm unit.
Supposedly, there must be no CO2 in the exhaust
line while calibrating the gas analyser but a very little
amount of CO2 could possibly exist. This also shows
the low amount of CO2 emitted by a hydrogen-fuelled
engine.
36
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: Fuel combustion
ACKNOWLEDGEMENT
The authors would like to thank the Ministry of Science, Technology
and Innovation for the initial grant under the IRPA project on CNGDI Engine as well as UTP for general support.
Abd Rashid Abd Aziz graduated with a
PhD in Mechanical Engineering (Thermofluid) from the University of Miami in 1995.
He is currently an Associate Professor and
the Research Head for Green Technology
(Solar, Hydrogen and Bio-fuels). He is also
the group leader for the Hybrid Vehicle
Cluster. He leads the Centre for Automotive
Research (CAR), which carried out several
research projects with funds from the Ministry of Science,
Technology and Innovation (MOSTI) and PETRONAS. His areas of
interest are in internal combustion engines, laser diagnostics, flow
visualisation, CFD, alternative fuels and hybrid powertrain.
VOLUME Six NUMBER two july - december 2008 PLATFORM
37
Technology Platform: fuel combustion
THE EFFECT OF DROPLETS ON BUOYANCY IN
VERY RICH ISO-OCTANE-AIR FLAMES
Shaharin A. Sulaiman and Malcolm Lawes1
Universiti Teknologi PETRONAS, 31750 Tronoh, Perak, Malaysia
1School of Mechanical Engineering, University of Leeds, UK
[email protected]
ABSTRACT
An experimental study is performed with the aim of investigating the effect of the presence of droplets in flames
of very rich iso-octane-air mixture under normal gravity. Experiments are conducted for initial pressures in the
range 100-160 kPa and initial temperatures 287-303 K at an equivalence ratio of 2.0. Iso-octane-air aerosols are
generated by expansion of the gaseous pre-mixture (condensation technique) to produce a homogeneously
distributed suspension of near mono-disperse fuel droplets. The droplet size varies with time during expansion;
hence the effect of droplet size in relation to the cellular structure of the flame was investigated by varying the
ignition timing. Flame propagation behavior was observed in a cylindrical vessel equipped with optical windows
by using schlieren photography. Local flame speeds were measured to assess the effect of buoyancy in gaseous
and aerosol flames. It was found that the presence of droplets resulted in a much earlier onset of instabilities, at a
rate faster than that taken for the buoyancy effect to take place. Flame instabilities, characterised by wrinkling and
cellular surface structure, increase the burning rate due to the associated increase in surface area. Consequently,
the presence of droplets resulted in a faster flame propagation rate than that displayed by a gaseous flame. The
mechanism of flame instabilities that caused a significant reduction of the buoyancy effect is discussed.
Keywords: buoyancy, combustion, droplets, flame, instabilities
INTRODUCTION
The combustion of clouds of fuel droplets is of
practical importance in engines, furnaces and also for
prevention of explosion and fire in the storage and
use of fuels. Theoretical [1] and experimental [2-4]
evidence suggests that flame propagation through
clouds of droplets, under certain circumstances, is
higher than that in a fully vapourised homogeneous
mixture. Even though this may be advantageous in
giving more rapid burning, its effects on emissions are
uncertain.
it is well established that the laminar burning rate
plays an important role in turbulent combustion [5].
Information on laminar burning velocity is sparse,
even for gaseous mixtures at conditions pertaining
to engines, which range from sub-atmospheric to
high pressure and temperature. Such data for fuel
sprays and for gas-liquid co-burning [6-8] are even
sparser than for gases. As a consequence, there is
little experimental data of a fundamental nature that
clearly demonstrates the similarities and differences
in burning rate, either laminar or turbulent, between
single and two-phase combustion.
Most combustion in engineering applications takes
place under turbulent conditions. Nevertheless,
In the present work, the influence of droplets in isooctane-air mixtures within the upper flammability
13th International Conference on Applied Mechanics and Mechanical Engineering, Egypt,
27-29 May 2008
38
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: Fuel combustion
Buoyancy driven flow velocity
Velocity resulted
expansion
from
gaseous
Gravity
the resulting
flame front
FLAME
FLAME
without buoyancy effect
with buoyancy effect
Figure 1. Illustration of the buoyancy effect [14] on spherical
flames. The arrows indicate the local velocities resulted by gas
expansion and buoyancy driven convection
limit (rich) was investigated. Such gaseous mixtures
are well known, from a number of previous works, for
example in [9-11], to experience the effect of buoyancy
or natural convection during combustion. Some of the
previous studies were performed in tubes. However,
Andrews and Bradley [12] suggested that the limit of
flammability obtained by the tube method would
be subject to the same sources of error [13] as would
be the burning velocity measurements using tubes,
mainly due to wall quenching. Thus studies in larger
tube diameters or in large explosion vessels have been
recommended.
Figure 1 shows an illustration of two centrally ignited
spherical flames to describe the effect of buoyancy
[14]. The open arrow represents the local velocity
which resulted from gas expansion during flame
propagation. The solid arrow, which points upward,
represents the local velocity caused by the buoyancy
or natural convection effect. The dashed lines show
the resulting flame front accounting for the net
velocity. With the absence of buoyancy effect, the
resulting flame would be spherical or circular when
viewed from the side. However, with the presence
of buoyancy effect, the resulting flame has a flatter
bottom surface as that illustrated in Figure 1.
(a)
SL
EV
28 litres
DL
VP
CV :
EV :
SL :
DL:
VP:
CV
23 litres
Combustion Vessel
Expansion Vessel
Supply Line
Discharge Line
Vacuum Pump
Orifice
Pipe
Valve
(b)
Figure 2. Aerosol combustion apparatus:
(a) photograph (b) schematic
EXPERIMENTAL APPARATUS
Figure 2 shows the photograph and schematic of
the aerosol combustion apparatus. The combustion
vessel, which essentially resembled a Wilson cloud
chamber [15], was a cylindrical vessel of 305 mm
diameter by 305 mm long (internal dimensions), with
a working volume of 23.2 litres. On both end plates of
the combustion vessel circular optical access windows
of 150 mm diameter were provided for characterisation
of aerosol and photography of flame propagation. To
initially mix the reactants four fans, driven by electric
motors, were mounted adjacent to the wall of the
VOLUME Six NUMBER two july - december 2008 PLATFORM
39
Technology Platform: fuel combustion
vessel. Two electrical heaters were attached to the
wall of the vessel to preheat the vessel and mixture
to 303 K. The expansion vessel, which has a volume
of 28 litres, was connected to the combustion vessel
by an interconnecting pipe through a port. The
vacuum pump, indicated in Figure 2 (a), was used to
evacuate the system and to remove particulates prior
to preparation of the mixture.
The aerosol mixtures were prepared by a condensation
technique, which generated near mono-dispersed
droplet suspensions. This was achieved by controlled
expansion of a gaseous fuel-air mixture from the
combustion vessel into the expansion vessel that
was pre-evacuated to less than 1 kPa. The expansion
caused a reduction in the pressure and temperature
of the mixture, which took it into the condensation
regime and caused droplets to be formed.
The characteristics of the generated aerosol were
calibrated by in-situ measurements of the temporal
distribution of pressure, temperature, and droplet
size and number, without combustion, with reference
to the time from start of expansion. The diameters
of individual droplets were measured using a Phase
Doppler Anemometer (PDA) system, from which the
droplet mean diameter, D10, was obtained. Since the
expansion took place over a period of several seconds
while combustion took place over less than 100 ms,
the far field values of D10 were assumed to be constant
during combustion.
The mean droplet diameter varied with time during
expansion; hence the effect of droplet size in relation
to the cellular structure of the flame was investigated
by varying the ignition timing. The iso-octane-air
mixture was ignited at the centre of the combustion
vessel by an electric spark of approximately 500 mJ.
The flame front was monitored through the vessel’s
windows by schlieren movies, which were recorded
using a high-speed digital camera at a rate of
1000 frames per second and with a resolution of 512 ×
512 pixels. The flame image was processed digitally by
using image-processing software to obtain the flame
radius. The velocity of the flame front, also known as
the stretched flame speed, S n , was obtained directly
40
from the measured flame front radius, r, by
Sn =
dr
dt
(1)
Similarly, the local flame speed is given by
SL =
dL
dt
(2)
where L is the distance between the local flame
front and the spark electrode as measured from the
schlieren image of the flame.
RESULTS AND DISCUSSION
Figure 3 shows the schlieren photographs of flames
at maximum viewable radius and the corresponding
contour plots at 2 ms intervals for gaseous and also
aerosol mixtures. The mixtures were initially at
equivalence ratio, φov, of 2.0, temperatures between
287 and 303 K, and pressures between 100 and
159 kPa. For the aerosol mixtures, the droplet mean
diameters, D10, were 5 µm and 13 µm. It must be
noted that the relatively small differences in pressure
and temperature between the conditions in the three
images have been shown, for gaseous flames, to have
little effect on the flame structure [16]. Hence, it was
assumed that the difference in the flame structure is
entirely due to the effects of droplets. In Figure 3(a),
the image is slightly darker than the others due to the
low exposure setting of the camera. The circular black
areas at the corners of the photographs represent
the region beyond the window of the combustion
vessel. The electrode holder and thermocouple are
clearly shown on the middle-right section in each
photograph.
It is shown in the schlieren image in Figure 3(a) that
for a gaseous flame, the upward distance of the flame
propagation is greater than the downward one, which
is a sign of the buoyancy effect as described for Figure
2. The upper surface of the gaseous flame is relatively
smoother and has fewer cells as compared to the lower
surface. In the contour plot for the gaseous flame in
Figure 3(a), larger spacing between flame contours
is shown for the upper propagation as compared to
the lower one. This suggests faster upward flame
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: Fuel combustion
(a) D10 = 0 µm (gaseous)
(b) D10 = 5 µm
(c) D10 = 13 µm
Figure 3. Schlieren images and contour plots (superimposition of edges) throughout propagation for iso-octane-air flames at φov = 2.0
and various droplet sizes. The time interval between each contour is 2 ms
propagation than the downward propagation.
Conversely, the difference between the leftward and
rightward propagations is shown to be small, the
rightward propagation being very close to the flame
radius. Hence it is shown that significant difference in
propagation rate occurs only in the vertical direction.
With the presence of droplets of 5 µm, it is shown in
Figure 3(b) that trend of vertical flame propagation is
almost similar to that for the gaseous flame in Figure
3(a). However the aerosol flame has more cells on
its surface than the gaseous one. Interestingly, with
bigger droplets (D10 = 13 µm), it is shown in Figure
3(c) that the resulting flame is highly cellular. The
contour plot of the flame shows faster burning rate
(large spacing between contours) and also smaller
difference between the upward and downward
propagation rate (more uniform contour spacing in
all directions) as compared to those in Figures 3 (a)
and (b). Hence it is suggested that with the presence
of large diameter droplets, the buoyancy effect
demonstrated in gaseous flames of rich mixtures is
overcome by instabilities and consequently faster
burning rate, such that there was less time available
for natural convection to be significant.
Figure 4 shows graphs displaying the temporal
variations of the flame radius and local vertical and
horizontal propagation distances (from the electrode)
for the gaseous iso-octane-air flame depicted in Figure
3(a) and for the aerosol flame in Figure 3(c), both at φov
= 2.0. Here the effect of the presence of large droplets
(D10 = 13 µm) is presented. The upward, downward,
leftward and rightward propagation distances
between the spark gap and the corresponding flame
edge were measured at 1 ms interval.
It is shown in the graph in Figure 4(a) that for the
gaseous flame, the upward propagation distance
of the flame propagation is always greater than
VOLUME Six NUMBER two july - december 2008 PLATFORM
41
Technology Platform: fuel combustion
70
U: upward
L: leftward
R: rightward
D: downward
r: schlieren radius
50
40
L
R
r
30
D
20
U
60
Distance (mm)
Distance (mm)
60
70
U
Propagation direction:
50
R
r
L
D
40
30
20
10
10
0
0
0
20
40
60
80
100
0
20
Time (ms)
(a) gaseous, D10 = 0 µm
40
60
80
100
Time (ms)
(b) aerosol, D10 = 13 µm
Figure 4. Flame propagation distance from spark electrode as a function of time for iso-octane-air mixtures at φov = 2.0. Also shown is
the corresponding direction of the flame propagation with respect to the spark electrode
that of the downward one. In addition, the upward
distance of the flame propagation increases at a
steady rate, whereas the downward one decelerates;
these indicate the effect of buoyancy force acting
on the hot flame kernel. The difference between the
leftward and rightward propagations is shown to be
small, although the rightward propagation distance
seems to be slightly different from the flame radius.
Obviously, the flame radius and horizontal propagation
distances are shown to be at approximately midway
between the upward and downward components.
The deceleration in the downward propagation is
only experienced by the gaseous flame, as shown in
Figure 4(a). Hence, the cellularity on the bottom half
of the gaseous flame in Figure 3(a) is probably caused
by hydrodynamic instabilities, as a result of an upward
flow of unburned gas which velocity exceeded the
flame speed at the base of the flame, as illustrated
in Figure 2. Conversely, the smoother upper surface
of the flame is probably due to flame stabilisation
as a result of high stretch rate. This occurs when the
expanding flame front propagates through a velocity
gradient in the unburned gas that is induced by the
upward, buoyant acceleration of hot products as
explained by Andrews and Bradley [12].
With the presence of large enough droplets (D10 = 13
µm), the aerosol flame burned faster than the gaseous
42
flame. This is shown in Figure 4(b), in which the aerosol
flame took approximately 60 ms to reach a radius of 50
mm, as compared to about 90 ms for the gaseous flame
to reach the same radius. This is very likely caused by
earlier instabilities in the aerosol flames, as depicted
in Figure 3, which promotes a faster burning rate due
to increase in the flame surface area. In relation to the
buoyancy, it is shown in Figure 4(b) that such effect
is absent, as implied by the gap between the upward
and downward flame propagation that is significantly
and consistently smaller as compared to that for
the gaseous flame. Furthermore, the aerosol flame
exhibits acceleration in the downward propagation as
compared to deceleration in the gaseous flame.
Figure 5 shows the vertical components of the flame
propagation distance from the spark electrode for
the gaseous flame and also the aerosol flames (D10
values of 5, 13, and 16 µm) at initial conditions similar
to those described for Figure 3. The negative values
indicate the downward flame propagation distances.
The propagation rates for the fine aerosol (D10 = 5 µm)
flames are shown in Figure 5 to be similar to those
for the gaseous flames, as indicated by their nearly
identical curve plots, particularly for the downward
propagation of the flame. In addition, the buoyancy
effect is evident by the greater values of the positive
distance as compared to the negative distance. It is
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: Fuel combustion
16 µm
Distance from spark electrode (mm)
60
13 µm
Upward
5 µm
0 µm
40
20
0
0
20
40
60
80
10 0
Time
(ms)
-2 0
0 µm
-4 0
5 µm
Downward
13 µm
16 µm
-6 0
Figure 5. Comparison of vertical flame propagation distance from spark electrode as a function of time for iso-octane-air mixtures at
φov = 2.0. Negative distances indicate downward flame propagation
clear in Figure 5 that the flames within aerosols of large
droplets (13 and 16 µm) propagate at a faster rate than
those of fine droplets (5 µm) and gaseous. For these
aerosol flames, the effect of buoyancy is not obvious.
Interestingly, it is shown that the curves for upward
propagation for all values of D10 are coincident for
approximately the first 25 ms of propagation; a similar
trend is observed for downward propagation up to
about 16 ms. This was probably because the buoyancy
is not yet in effect during those periods of initial flame
kernel growth.
Figure 6 shows graphs of variation in the local flame
speed (deduced by time derivative of the graphs
in Figure 4) with time from the start of ignition.
The speed for the upward propagating flame is
represented by the diamond markers, and the bottom
one by the square markers. The circle and triangle
markers represent the rightward and leftward flames
respectively. The gaseous flame is shown in Figure
6(a) to propagate faster in the upward component by
about 0.4 m/s, as compared to that of the downward
component, which also decreases towards a near zero
value throughout propagation. A negative downward
component, if were to occur, would implicate an
upward flow of unburned gas at the central base of
the flame; such cases were reported elsewhere; e.g.
in [9], but this is beyond the scope of the present
work. The sideway components of the flame speed
are shown in Figure 6(a) to be similar, which suggest
the independencies of these components from the
natural convection effect in the gaseous flame. With
droplets (D10 = 13 µm), it is shown in Figure 6(b) that all
the components of flame speed nearly coincide with
each other, indicating a more uniform distribution of
flame speed throughout the flame surface, and hence
evident the absence of the natural convection effect.
However, after about 35 ms from the start of ignition,
the curve for the upward component of the flame
started to burn at a significantly faster rate than the
other components. The reason for this is not clear and
thus further investigation is required.
VOLUME Six NUMBER two july - december 2008 PLATFORM
43
Technology Platform: fuel combustion
1.8
Upward
Downward
Leftward
Rightward
Local Flame Speed (m/s)
.
1.6
1.4
1.2
1.0
0.8
0.6
0.4
0.2
0.0
0
20
40
60
80
100
Time (ms)
a) gaseous, D10 = 0 µm
1.8
1.6
Upward
Downward
Leftward
Rightward
Local Flame Speed (m/s).
1.4
1.2
1.0
0.8
0.6
0.4
0.2
0.0
0
(b) aerosol, D10 = 13 µm
20
40
60
80
100
Time (ms)
Figure 6. Comparison of vertical flame propagation distance from spark electrode as a function of time air for iso-octane-air mixtures at
φov = 2.0 for (a) D10 = 0 µm (gaseous) and (b) D10 = 13 µm
44
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: Fuel combustion
The mechanism of flame instabilities, which caused
increased cellularity and insignificance of the
buoyancy effect, in aerosol flames is probably related
to the heat loss from the flame and local rapid
expansion through droplet evaporation. Although
droplet evaporation can also cause high gradients in
the mixture strength (variations in local equivalence
ratio), which might have an effect on the flame, this
was negated experimentally [17] using water aerosol
in propane-air mixtures. In another study [18] using a
rig similar to that of the present work, it was shown
that the presence of 30 µm diameter hollow spherical
glass beads (no evaporation) in a gaseous iso-octaneair mixture did not alter the smooth characteristics of
the flame structure as well as the burning rate. Thus it
is evident that the presence of droplets probably plays
an important role in the introduction of instabilities
due to evaporation.
propensity to instability results in the burning rate of
aerosol mixtures being faster than those in the gaseous
phase at similar conditions. This is so, although the
fundamental unstretched laminar burning velocity is
probably unchanged by the presence of droplets.
REFERENCES
[1]
J. B. Greenberg, “Propagation and Extinction of an Unsteady
Spherical Spray Flame Front,” Combust. Theory Modelling,
vol. 7, pp. 163-174, 2003
[2]
D. R. Ballal and A. H. Lefebvre, “Flame Propagation in
Heterogeneous Mixtures of Fuel Droplets, Fuel Vapor and
Air,” Proc. Combust. Inst., 1981
[3]
G. D. Myers and A. H. Lefebvre, “Propagation in
Heterogeneous Mixtures of Fuel Drops and Air,” Combustion
and Flame, vol. 66, pp. 193-210, 1986
[4]
G. A. Richards and A. H. Lefebvre, “Turbulent Flame Speeds
of Hydrocarbon Fuel Droplets in Air,” Combustion and
Flame, vol. 78, pp. 299-307, 1989
[5]
D. Bradley, A. K. C. Lau, and M. Lawes, “Flame Stretch Rate as
a Determinant of Turbulent Burning Velocity,” Phil. Trans. R.
Soc. Series A, vol. 338, pp. 359, 1992.
[6]
Y. Mizutani and A. Nakajima, “Combustion of Fuel VapourDrop-Air Systems: Part I, Open Burner Flames”, Combustion
and Flame, vol. 20, pp. 343-350, 1973
[7]
Y. Mizutani and A. Nakajima, “Combustion of Fuel VapourDrop-Air Systems: Part II, Spherical Flames in a Vessel”,
Combustion and Flame, vol. 21, pp. 351-357, 1973
[8]
F. Akamatsu, K. Nakabe, Y. Mizutani, M. Katsuki, and T. Tabata,
“Structure of Spark-Ignited Spherical Flames Propagating
in a Droplet Cloud”, in Developments in Laser Techniques
and Applications to Fluid Mechanics, R. J. Adrian, Ed. Berlin:
Springer-Verlag, 1996, pp. 212-223
[9]
H. F. Coward and F. Brinsley, “Dilution Limits of Inflammability
of Gaseous Mixtures”, Journal of Chemical Society
Transaction (London), vol. 105, pp. 1859-1885, 1914
CONCLUSION
The effects of the presence of near mono-dispersed
droplets in flames of very rich iso-octane-air mixture
(φov = 2.0) were investigated experimentally in a
spherical explosion vessel at near atmospheric
conditions. The fuel droplets, which were in the form
of aerosols/vapour, were generated by condensation
of the gaseous pre-mixture through expansion and this
resulted in a homogeneously distributed suspension
of near mono-disperse fuel droplets. The effects of
droplet size in relation to the structure of the flame
surface and to the burning rate were investigated by
varying the ignition timing, as the droplet size varied
with time during expansion. Observations of the
gaseous flame using schlieren photography through
the vessel’s windows revealed the buoyancy effect,
with distinct differences in flame surface structure
and local burning rates between the upper and
lower halves of the flame, similar to those described
in previous studies. The presence of fine droplets
(5 µm) did not cause significant change with respect
to the gaseous flame in terms of the buoyancy effect,
flame structure and burning rate. However, with
larger droplets (13 µm) the flame became fully cellular
at a faster rate and more importantly the effect of
buoyancy was significantly reduced. The increased
[10] O. C. d. C. Ellis, “Flame Movements in Gaseous Mixtures”,
Fuel, vol. 7, pp. 195-205, 1928
[11] I. Liebman, J. Corry, and H. E. Perlee, “Dynamics of Flame
Propagation through Layered Methane-Air Mixtures”,
Combustion Science and Technology, vol. 2, pp. 365, 1971
[12] G. E. Andrews and D. Bradley, “Limits of Flammability
and Natural Convection for Methane-Air Mixtures”, 14th
Symposium (International) on Combustion, 1973
[13] G. E. Andrews and D. Bradley, “Determination of Burning
Velocities, A Critical Review”, Combustion and Flame, vol.
18, pp. 133-153, 1972
[14] D. Bradley, Personal Communications, 2006
VOLUME Six NUMBER two july - december 2008 PLATFORM
45
Technology Platform: fuel combustion
[15] C. T. R. Wilson, “Condensation of water vapour in the
presence of dust-free air and other gases”, in Proceedings of
the Royal Society of London, 1897
[16] D. Bradley, P. H. Gaskell, and X. J. Gu, “Burning Velocities,
Markstein Lengths and Flame Quenching for Spherical
Methane-Air Flames: A Computational Study”, Combustion
and Flame, vol. 104, pp. 176-198, 1996
[17] F. Atzler, M. Lawes, S. A. Sulaiman, and R. Woolley, “Effects of
Droplets on the Flame speeds of Laminar Iso-Octane and
Air Aerosols”, ICLASS 2006, Kyoto, Japan, 2006
[18] F. Atzler, “Fundamental Studies of Aerosol Combustion”,
Department of Mechanical Engineering, University of
Leeds, 1999
46
Shaharin Anwar Sulaiman graduated in
1993 with a BSc in Mechnical Engineering
from Iowa State University. He earned his
MSc in Thermal Power and Fluids
Engineering from UMIST in 2000, and PhD
in Combustion from the University of Leeds
in 2006. During his early years as a graduate,
he worked as a Mechanical and Electrical
(M&E) Project Engineer in Syarikat
Pembenaan Yeoh Tiong Lay (YTL) for five years. His research
interests include combustion, sprays and atomisation, airconditioning and ventilation, and biomass energy. He joined UTP
in 1998 as a tutor. At present he is a Senior Lecturer in the
Mechanical Engineering programme and also the Programme
Manager for MSc in Asset Management and Maintenance.
Certified as a Professional Engineer with the Board of Engineers,
Malaysia. He is also a Corporate member of the Institution of
Engineers Malaysia.
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: SYSTEM OPTIMISATION
Anaerobic Co-Digestion of kitchen waste
and sewage sludge for producing biogas
Amirhossein Malakahmad*, Noor Ezlin Ahmad Basri1, Sharom Md Zain1
*Universiti Teknologi PETRONAS, 31750 Tronoh, Perak Darul Ridzuan, Malaysia.
1Universiti Kebangsaan Malaysia, 43600 UKM, Bangi, Selangor, Malaysia.
*[email protected]
ABSTRACT
In this paper, an attempt was made to present the results of some experiments conducted on anaerobic digesters
to make comparative study of the biogas generation capacity of the mixture of organic fractions of municipal
solid waste from kitchen waste and sewage sludge in different compositions. Batch digestion of samples with
various percentage of kitchen waste and sewage sludge was carried out under controlled temperature 35 °C
and pH7 conditions for 15 days for each experiment. In all experiments the content of total solid and volatile
solid, pH, Kjeldahl nitrogen, chemical oxygen demand, biogas productivity and the content of biogas were
measured. The results obtained showed that biogas productivity varied between 4.6 and 59.7 ml depending on
the composition of each component in the sample which were added to the digesters. The bioprocess efficiency
was observed to be 7.5% - 70.5% for total solid, 51.2% - 81.0% for volatile solid and 8.3% - 43.8% for COD. The
overall effluent chemical oxygen demand concentration indicated that it should be treated before using for
other applications. From the study results, it is evident that the second bioreactor with 75% of kitchen waste and
25% of sewage sludge produced the maximum quantity of methane gas as compared to other bioreactors.
Keywords: kitchen waste, sewage sludge, biogas production
INTRODUCTION
material recycling, energy recovery and landfill [1].
Malaysia, with a population of over 25 million, generates
16 000 tones of domestic waste daily. At present, the
per capita generation of solid waste in Malaysia varies
from 0.45 to 1.44 kg/day depending on the economic
status of an area. There are now 168 disposal sites but
only 7 are sanitary landfills. The rest are open dumps
and about 80% of these dumps have filled up to the
brim and have to be closed in the near future. The
Malaysian government is introducing a new law on
solid waste management and also drafting a Strategic
Plan for Solid Waste Management in Peninsular
Malaysia. The principal processes options available
and being recognised as hierarchy for integrated
waste management are: waste minimisation, reuse,
Municipal solid waste (MSW) contains an easily
biodegradable organic fraction (OF) of up to 40%.
Conventional MSW management has been primarily
disposal by land filling. Sewage sludge is characterised
by high content of organic compounds and this is the
cause of its putrescibility. Therefore, sludge before
landfill disposal or agricultural application should
undergo chemical and hygienic stabilisation. One
possible method of stabilisation and hygienisation
involves methane fermentation [2].
The anaerobic co-digestion of sewage sludge with
organic fraction of municipal solid waste (OFMSW)
seems to be especially attractive [3]. The feasibility of
This paper was presented at the 2nd International Conference on Environmental Management, Bangi,
13 -14 September 2004
VOLUME Six NUMBER two july - december 2008 PLATFORM
47
Technology Platform: SYSTEM OPTIMISATION
anaerobic co-digestion of waste activated sludge and
a simulated OFMSW was examined by Poggi-Varaldo
and Olesz-kiewicz [4]. The benefits of co-digestion
include: dilution of potential toxic compounds,
improved balance of nutrients, synergistic effects of
microorganisms, increased load of biodegradable
organic matter and better biogas yield. Additional
advantages include hygienic stabilisation, increased
digestion rate, etc. during methane fermentation the
two main processes that occur are:
i. Acidogenic digestion with the production of
volatile fatty acid; and
ii. The volatile fatty acids are converted into CH4 and
CO2.
In batch operation the digester is filled completely
with organic matter and seed inoculums, sealed, and
the process of decomposition is allowed to proceed
for a long time until gas production is decreased to a
low rate (duration of process varies based on regional
variation of temperature, type of substrate, etc.).
Then it is unloaded, leaving 10-20 percent as seed,
then reloaded and the operation continues. In this
type of operation the gas production is expected to
be unsteady and the production rate is expected to
vary from high to low. Digestion failures due to shock
load are not uncommon. This mode of operation,
however, is suitable for handling large quantities of
organic matter in remote areas. It may need separate
gasholders if a steady supply of gas is desired [5].
Callaghan et al. worked on co-digestion of waste
organic solids which gave high cumulative production
of methane [6].
kitchen waste and 50% sewage sludge, in the fourth
run, mixture of 25% kitchen waste and 75% sewage
sludge was used. The fifth experiment was conducted
with only sewage sludge.
In all experiments total solid, volatile solid, pH, Kjeldahl
nitrogen and chemical oxygen demand for initial and
final properties of samples were determined. Biogas
productivity and the content of biogas were also
measured. All analytical procedures were performed
in accordance with Standard Methods [7].
Results and discussion
i. Variation in pH value throughout the
experiments
As shown in Figure 1, from the graph plotted, the
pH variation could be categorised into 3 main zones.
The first zone started from first day till fourth day,
which showed a drastic drop of the pH. This is due
to the high development rate of volatile fatty acids
by microorganisms. The pH is maintained at neutral
with the addition of sodium hydroxide solution. The
second zone started from the fifth till the twelfth day
of experiment. In the second zone, the pH was in the
range of 6.9 to 7.3. This is due to the development
of CO3HNH4 from CO2 and NH3, which were produced
during the anaerobic process.
The percentage of CO3HNH4 had caused the increase
of alkalinity of the samples. Due to this, any differences
Experimental
7
6
pH
Batch digestion of samples was carried out under
controlled temperature 35 °C and pH7 conditions for
15 days for each experiment. All five samples were
fed into a 1 L bioreactor operated under mesophilic
condition.
8
5
Sample 1
4
Sample 2
Sample 3
3
Sample 4
Sample 5
In the first experiment, 100% kitchen waste was used
while the second experiment was conducted with the
mixture of kitchen waste (75%) and sewage sludge
(25%). The third experiment was conducted with 50%
48
2
1
2
3
4
5
6
7
8
9
Time (day)
Figure 1
PLATFORM VOLUME Six NUMBER TWO july - december 2008
10
11
12
13
14
15
16
Technology Platform: SYSTEM OPTIMISATION
in the volatile fatty acid content did not affect the pH
value. The third zone started on the thirteenth till
the last day of the experiment. In this zone, it was
found that the pH value of the samples started to
increase. This is due to the development of CO3HNH4
that still continues, but no more volatile fatty acid was
produced.
ii. Biogas production
• The production of cumulative biogas
Figure 2 shows the production of cumulative biogas
for all the samples. It was found that the second
sample with the composition of 75% kitchen waste
and 25% activated sludge produced the highest
quantity of biogas, which was 59.7 ml. This was
followed by the first sample (100% kitchen waste),
then fourth sample (25% kitchen waste and 75%
activated sludge), then the third sample (50% kitchen
waste and 50% activated sludge), and lastly the fifth
sample (100% activated sludge). The productions of
the biogas of the respective samples were 47.1 ml,
22.3 ml, 8.4 ml and 4.6 ml. The fifth sample produced
the least biogas; this is consistent with the literature
data by Cecchi et al., (1998) and Hamzawi et al., (1998)
that showed the production of the cumulative biogas
is high when organic components that are easily
biodegradable in the sample are higher. According to
Schmidell (1986), the anaerobic digestion process for
MSW alone is possible but will produce less biogas as
compared to a mixture of MSW and activated sludge.
This is due to the production of volatile fatty acids by
microorganisms is more likely to accumulate rather
than to release biogas.
An increment of 5% of activated sludge is enough
to reduce the accumulation of volatile fatty acid and
release more biogas. From Figure 2 for the first and
second samples, the results comply with the Schmdell
theory. Samples three and four that had different
composition of kitchen waste and activated sludge
produced less biogas due to the composition of
activated sludge that has unsuitable C:N ratio for the
anaerobic digestion process.
•
The biogas production rate
The rate of biogas production for every sample is
shown in Figure 3. It was found that the production of
biogas for the first sample was the slowest that started
from the fourth day of experiment and reached the
highest quantity on the eighth day. On the other hand
the fifth sample started to produce biogas the earliest
on the second day and reached the highest amount
on the fifth day. The production rate of biogas for
samples two and three occurred on the seventh day of
the experiment while the fourth sample occurred on
the sixth day. Therefore, it can be concluded that the
results obtained are consistent with the research done
by Cecchi et al. [8], which stated that the production of
biogas is slower for high organic loading as compared
to a lower organic loading.
Table 1 summarises the amount of total suspended
solid (SS), volatile suspended solid (VSS), alkalinity,
30
Sample 1
Sample 1
60
Sample 3
Sample 3
50
Sample 2
25
Sample 2
Biogas production rate (ml/day)
Production of cumulative biogas (ml)
70
Sample 4
Sample 5
40
30
20
10
Sample 4
20
Sample 5
15
10
5
0
0
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
1
Figure 2
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
Time (day)
Time (day)
Figure 3
VOLUME Six NUMBER two july - december 2008 PLATFORM
49
Technology Platform: SYSTEM OPTIMISATION
kejldahl nitrogen, pH, and COD before and after
treatment in all five experiments. According to the
results, the bioprocess efficiency was observed to be
7.5% - 70.5% for total solid, 51.2% - 81% for volatile
solid and 8.3% - 43.8% for COD. The first bioreactor is
most efficient in treating the volatile solid component,
which achieved efficiency of 81.0%.
Conclusions
Five experiments were conducted under mesophilic
conditions in batch bioreactor for 75 days. Five
different kinds of feedstock were loaded into the
reactor. It was found that the cumulative biogas
production increased, when the mixture kitchen
waste and activated sludge was used.
However, the highest value of methane production
was for sample 2 (75% kitchen waste and 25%
activated sludge), which produced 59.7 ml. For
the rate of biogas production the situation was the
same and the best result was for sample 2, after the
samples 1, 4, and 3 were settled respectively. The 5th
sample produced the least biogas. The anaerobic
co-digestion of kitchen waste and activated sludge
were demonstrated to be an attractive method for
environmental protection and energy savings, but it is
clear that applying better equipment and adjustment
of conditions could give more reasonable results.
Reference
[1]
Mageswari, S., “GIAI global meeting”. Penang, Malaysia. 1721 March 2003
[2]
Sosnowski, P., Wieczorek, A., Ledakowicz, S., 2003. “Anaerobic
co-digestion of sewage sludge and organic fraction
of municipal solid wastes”. Advance in environmental
researches 7, 609-616
[3]
[4]
50
Table 1. The value of measured parameters before and after
treatment
Parameter
Before
After
treatment treatment
Sample I
SS (%)
5.00
3.29
VSS (%)
42.14
8.00
N (mg/L N-NH3)
0.40
0.55
pH
7.00
7.36
COD (mg/L)
4864
3520
SS (%)
5.10
4.72
VSS (%)
38.30
7.50
N (mg/L N-NH3)
0.29
0.47
pH
7.00
7.31
COD (mg/L)
4000
3600
SS (%)
6.44
1.90
VSS (%)
32.55
8.80
N (mg/L N-NH3)
0.24
0.45
pH
7.01
7.27
COD (mg/L)
2800
2560
SS (%)
2.45
1.40
VSS (%)
27.52
6.69
N (mg/L N-NH3)
0.23
0.30
pH
7.00
7.50
COD (mg/L)
2400
2200
Sample II
Sample III
Sample IV
Hamzawi, N., Kennedy, K.J., Mc Lean, D.D., 1998. “Technical
feasibility of anaerobic co-digestion of sewage sludge and
municipal solid waste”. Environ. Technol. 19, 993-1003
Sample V
SS (%)
0.27
0.19
Poggi-Varaldo, H. M., Olesz-kiewicz, J. A., 1992. “Anaerobic
co-composting of municipal solid waste and waste sludge
at high total solid level”. Environ technol. 13, 409-421
VSS (%)
0.41
0.20
N (mg/L N-NH3)
0.02
0.10
pH
7.00
7.59
COD (mg/L)
320
180
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: SYSTEM OPTIMISATION
[5]
Polprasert, C., 1996. “Organic waste recycling, technology
and management”. Second edition. Chichester: John Wiley
and sons
[6]
Callaghan F. J., Wase, D. A. J., Thayanithy, K., Forster, C. F.,
1998. “Co-digestion of waste organic solids: batch study”.
Bioresearch technology 67, 117-122
[7]
Standard methods for water and wastewater treatment
18th edition (1992)
[8]
Cecchi, F., Pavan, P., Musacco, A., Mata-Alvarez, J., Sans, C.,
Ballin, E., 1992. “Comparison between thermophilic and
mesophilic digestion of sewage sludge coming from urban
wastewater plants”. Ingegneria Sanitaria Ambientale 40, 2532
Amirhossein Malakahmad is an academic
staff at Universiti Teknologi PETRONAS. He
graduated with BEng in Chemical
Engineering in 1999 from Islamic Azad
University, Tehran, Iran. He completed his
MSc in 2002 in Environmental Engineering
from the same university. In 2006, he
received his PhD from the National
University of Malaysia, UKM for his research
on an application of zero-waste anaerobic baffled reactor to
produce biogas from solid waste. He joined Universiti Teknologi
PETRONAS in August 2007. His research interests are in water and
wastewater treatment and solid waste engineering.
VOLUME Six NUMBER two july - december 2008 PLATFORM
51
Technology Platform: SYSTEM OPTIMISATION
On-line At-Risk Behaviour Analysis
and Improvement System (e-ARBAIS)
Azmi Mohd Shariff* and Tan Sew Keng
Universiti Teknologi PETRONAS, 31750 Tronoh, Perak Darul Ridzuan, Malaysia.
*[email protected]
Abstract
Behaviour Based Safety (BBS) is a programme that has been implemented in many organisations to identify at-risk
behaviour and to reduce injury rate of their workers. The effectiveness of the BBS was proven with many companies
recorded high percentage of reduction of injury rate especially during the first year of implementation. However,
the BBS process could be very labour intensive. It requires many observers to make the process effective. Very
much effort was required to train the employees to become the observers. Many organisations which attempted
to obtain the benefits of BBS did not sustain comprehensive participation required in BBS related activities.
With this drawback, it calls for a simplified process that could achieve the same result as BBS. This study was
intended to establish an alternative to the BBS, termed as On-line At-Risk Behaviour Analysis and Improvement
System (e-ARBAIS). The e-ARBAIS utilises computer technology to play a role in making the routine observation
process more sustainable and hence instilling the habitual awareness through the cognitive psychology effect.
A database was set up with the pre-programmed questions regarding at-risk behaviours in the organisation.
The employees then utilised the database to feedback their observations. Through this process, the traditionally
tedious observations by trained observers as required in BBS were now done naturally by all respondents. From
the collective feedback, at-risk behaviours can be easily identified in the organisation. The HSE committee within
the organisation can thus, take the appropriate action by reinforcing the safety regulations or safe practices to
correct all the unsafe behaviours either by changing the design (“Hard-ware”), system (“Soft-ware”) or training
the employee (“Human-ware”). This paper introduces the concept, framework and methodology of e-ARBAIS.
A case study was conducted in X Company (not a true name as the permission to use their real name was not
given). A prototype computer program based on e-ARBAIS was developed using Microsoft Excel named as
“1-min Observation” programme. A preliminary result based on one-month data collection is highlighted in
this paper. Based on this preliminary result, “1-min Observation” programme has received positive feedback
from the management and employees of Company X. It was done with very small resources and thus saved
time and money compared to traditional BBS technique. The e-ARBAIS concept is workable and practical since it
is easy to implement, collect data and correct unsafe behaviour. Some recommendations by the employees of
Company X were presented in this paper to further improve the “1-min Observation” programme. The project
at Company X is still in progress in order to see a long term impact of this programme. The e-ARBAIS has shown
its potential to reduce injury in the organisation if implemented with a thorough plan and strong commitment
from all levels.
Keywords: behaviour based safety, at-risk behaviour, human factor, safety programme, injury rate reduction
This paper was published in the Journal of Loss Prevention in the Process Industries,
doi:10.1016/j.jlp.2007.11.007 /(2008)
52
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: SYSTEM OPTIMISATION
Introduction
Behaviour based safety (BBS) was first established
by B. F. Skinner in the 1930s (Skinner, 1938). He was
a psychologist who developed a systematic approach
called behaviour analysis to increase safe behaviours,
reduce risky behaviours and prevent accidental
injury at work and on the road. This approach was
later known as applied behaviour analysis (Hayes,
2000). Behaviour study was important because H. W.
Heinrich, a workplace safety pioneer, reported that
out of 550 000 accidents, only 10% was caused by
unsafe working conditions, another 88% was caused
by worker’s unsafe actions (SCF, 2004). A “Workplace
Attitude Study” conducted by Missouri Employers
Mutual Insurance (MEM) which was published in
Occupational Hazards (September 2003) revealed that
64.1% of Americans thought that a workplace accident
would never happen to them. 53.4% believed that the
probability was very low for a work injury that could
cause them to become permanently disabled (SCF,
2004). This showed that people generally perceived
that there was a low risk of injury possibility in a
workplace. This showed that accidents could happen
if workers continue to work with at-risk behaviour and
perceived it was safe to do so. The human toll of unsafe
behaviour was high. According to the U.S. Bureau of
Labour Statistics, unintentional injury was the leading
cause of death to people aged 44 and under. In 2001,
private industry had more than 5.2 million non-fatal
accidents and injuries, with more than 5 000 fatal
injuries. Behaviour-based safety programmes that
target and document behaviour changes indeed save
lives, money and productivity (APA, 2003).
Behaviour is an “upstream” approach to safety. It
focuses on the “at-risk behaviour” that might produce
an accident or near miss rather than trying to correct
a problem after an accident or occurrence. The
behaviour-based aim then, is to change the mindset
of an employee by hopefully making safety a priority
in the employee’s mind (Schatz, 2003).
However, it was noted to many that not all organisations
had successful experience in implementing the BBS as
the others did (Geller, 2002). Over years, some safety
professionals had started to develop alternatives to
the BBS programme, i.e. people-based safety, ProAct
Safety and Value-based Safety. It was desirable to
develop another alternative to the BBS programme
via the help of computer technology.
The e-ARBAIS Concept
The e-ARBAIS programme is meant to provide
alternative solutions to certain limitations of BBS as
mentioned below.
a. Prevent coyness in direct feedback with
computer interface
Problem arises when employees dare not approach
the peers to give feedback directly (Gilmore et al.,
2002). The e-ARBAIS programme provided another
channel for the peers to communicate. The peers
may now give feedback on their observations to the
database and publish the feedback via the computer.
This helps to reduce the problem of coyness through
direct feedback with peers.
b. Inculcate safety culture with e-ARBAIS
The e-ARBAIS utilises computer software to prompt
the employees if they observe any unsafe behaviour
relating to the topic asked. For instance, the
e-ARBAIS database could have a question like, “Did
you see anybody drive faster than 20 km/h today?”
The employees would be reminded to observe
occurrence around them naturally without bringing
the checklist. The e-ARBAIS questions that would be
prompted regularly may also serve as the checklist
in the ordinary BBS process. However, instead of
focusing on many items during one observation, the
questions would require the employees to respond
to certain particular areas in a day. Different topics
would be asked everyday. This will eventually instil a
psychological effect in which people are “reminded”
on the safety rules and regulations. The cultivating
of habitual awareness is always the heart of designing
the e-ARBAIS. Only when it becomes habitual, the
safety culture could be inculcated.
VOLUME Six NUMBER two july - december 2008 PLATFORM
53
Technology Platform: SYSTEM OPTIMISATION
As employees perform observations, they come
to recognise any discrepancies between their own
behaviour and what is considered safe, and they
begin to adopt safe practices more consistently.
McSween said “We have created a process where they
raise their personal standards” (Minter, 2004). This is
the objective of this whole e-ARBAIS – to inculcate the
safety culture in organisations.
c. Tackling slow data collecting and analysis with
IT
Questions on observations are then being repeated
randomly. All the feedbacks are collected in the
database and would be analysed by the HSE committee
on a regular basis. When the HSE committee analyse
the data, they would be able to identify the high
risk issues faced by the employees. For example, if
the speeding behaviour of the employees remains
high in the statistics after a few times being asked
through the software, it implies that many people
are speeding in the plant and refuse to change their
behaviour. From there, the HSE committee could
provide some recommendations such as building a
few road humps in the plant, doing spot checks and
issuing warnings to those who speed. The action item
could also be derived from an established method
such as ABC analysis or ABC model (Minshall, 1977).
The data analysed would highlight the areas in which
the HSE committee needs to focus on and to provide
the solution for improvement. This would also help
the committee to identify if the unsafe behaviours are
due to
• "Hard-ware" problem like inadequate structural
safety design,
• "Soft-ware" problem like poor system
implementation or obsolete operating procedure,
or
Set up of e-ARBAIS Database
Briefing to End Users on
e-ARBAIS
Feedback to peers
(optional)
Observations of At-Risk Behavior
at Workplace
On-line Feedback to Database
Analysis of At-Risk Behavior
ABC Analysis
(optional)
Review by HSE Committee
Action Plan to Correct At-Risk
Behavior
Figure 1. Framework of e-ARBAIS
54
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Interview
(optional)
Technology Platform: SYSTEM OPTIMISATION
Table 1. The comparison of ordinary BBS and e-ARBAIS
Element
Ordinary BBS
Training
Training given to observers on how to No training needed. All will participate in
define unsafe behaviour and how to provide the observations. Only briefing on what is
feedback.
e-ARBAIS and how it works.
Checklist
Checklist must be used to go through all No checklist is used. Checklist is built in the
items and see which does not comply.
database as pre-programmed questions.
Observation
frequency
Observers are required to make certain Observation is done on daily basis or flexible
observations in a certain period, i.e. 1 adjustment to frequency can be made.
observation a week.
Cost
Additional cost for training and printing Minimum cost since training and checklist
checklist.
are not required.
Feedback
Feedback is given directly when observation
completes, whether it is positive or negative.
Results of observation need to be reported
to HSE committee for further analysis.
Communication
The analysed results normally can only be Feedback is recorded in database which
accessed by the HSE committee. Employees everyone could access to see what unsafe
generally not being communicated on the behaviour is observed.
overall performance of the observation.
Involvement
Only those who are trained will be involved All will be involved in the observation since
in the observations. To involve all, much the observation is done naturally without
training is required.
focusing on only one activity.
Management
commitment
Management commitment determines if Database displays the management
the process will be successful. Most BBS fail participation
and
thus
motivates
due to poor management commitment.
management to further commit and
improve the programme.
•
e-ARBAIS
"Human-ware" problem like employees’ risky
behaviour.
Feedback can be given either by face to face
or through the database to prevent “sick
feeling” with peers. Feedback is displayed
directly in the database. HSE committee
uses the same set of data for further action.
Framework of e-ARBAIS
The framework of e-ARBAIS is given in Figure 1.
d. Reduce intensive labour and high cost with
e-ARBAIS 24 hour functionality
The advantage of e-ARBAIS is that the observation
can be done 24 hours a day and 7 days a week with
minimum man-hour needed. It is all operated by the
computer. Thus, resulting in higher efficiency and
cost saving.
The differences between e-ARBAIS to the ordinary
BBS programme are given in Table 1.
To start the e-ARBAIS program, a database first
needs to be set up. A systematic approach can be
established to develop the questions in the database.
The database consists of pre-programmed questions
that can be based on the past incident records in the
organization to focus on the problematic area, or it can
be the near miss cases. It can also be the questions that
purely derived from the BBS checklist alone. Basically,
to effectively use the e-ARBAIS program, the questions
must be custom designed for each company. The
discussion needs to be done with HSE committee on
VOLUME Six NUMBER two july - december 2008 PLATFORM
55
Technology Platform: SYSTEM OPTIMISATION
the development and selection of the questions and
agreed by the management of the company.
After that, the end users needed to be briefed
about the e-ARBAIS and how the program will be
implemented. The end users were not required to
observe any specific activities to ensure that e-ARBAIS
is naturally done according to the pre-programmed
questions in the database. The end users only needed
Table 2. The questions in “1-min Observation” database
1. Did you see people NOT using handrail when
travelling up and down stairs?
2. Did you see employee driving faster than
20km/h in the plant area?
3. Did you experience back pain after work?
4. Did you see any reckless forklift driver?
5. Did you see people wearing glasses instead of
safety glasses in the plant?
6. Did you see people lifting thing in improper
position?
7. Did you see people NOT wearing ear plug in
noisy area?
8. Did you see anyone working at height with
falling hazard due to improper PPE/position?
9. Did you see people working (in the plant/lab/
packaging) NOT wearing safety glass?
10. Did you see any leak in the plant but NOT
barricaded?
11. Did you see any area (office or plant) unclean
and expose to tripping hazard?
12. Did you see people using hand phone at
restricted area?
13. Did you see people NOT looking in the direction
that they are walking (eyes NOT on path)?
14. Did you see people NOT following 100% of the
procedure when doing work?
15. Did you see people working without hand
glove in the plant area?
16. Did you see people walking/working inside
the maintenance workshop yellow line area
without minimum PPE required?
56
to be more aware of at-risk behaviour observation
by their colleagues when they are doing daily work.
At any occasion, they could also give their feedback
directly to the peers. To those who are afraid that their
feedback could cause ill-feeling, they have the option
of giving their feedback to the database. The database
would calculate the analysis automatically and publish
the result online based on the feedback received.
The employees are indirectly reminded by the online
analysis and feedback on the unsafe behaviours from
this exercise and this could change their behaviour
eventually to avoid similar observations in the
future. This could be achieved through the cognitive
psychology effect.
With the analysis done by the database, the HSE
committee could use those data and discuss them
in their regular meeting directly to correct the at-risk
behaviours that frequently occur. The committee could
apply the ABC analysis (Minshall, 1977) to understand
the behaviours. Optionally, the HSE committee could
also conduct some interviews with the identified
employees to further understand why they take risk
when performing their tasks. With that, the necessary
action plan could be established to rectify the at-risk
behaviours that are contributed by factors such as
“hard-ware”, “soft-ware” or “human-ware” mentioned
above.
The questions in the pre-programme database need
to re-visit time to time and should be updated with
any new potential at-risk behaviour as necessary.
Case Study
A case study using e-ARBAIS concept was implemented
in Company X for a month. The program was termed
as “1-min Observation”. One of the important tools
of this study was to use IT (information technology)
to help on the observation processes. The database
was developed on Microsoft Excel spreadsheet. One
master list that consists of 16 questions was generated
in the database based on the previous frequent atrisk behaviour identified and reported as shown in
Table 2. The case study was run for a month and each
question was repeated four times to observe any
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: SYSTEM OPTIMISATION
Start
User open database
Database would show today's date and search
for questions which matched with today's date.
Database would prompt the 2 questions which
matched with today's date into the front page
User to select "yes" or "no" for both questions
and department name
User clicked "submit".
Data was captured in database. Data collected
consist of date, user ID, answer for 1st question,
answer for 2nd question, and department.
Database would
count the
participation in a
day based on
departments. The
data was
transferred as chart
and would be
shown in the page
of "participation".
Database would count the
percentage of total unsafe
behavior observed in a day.
The data was transferred to
chart and would be shown in
the page of "unsafe
behavior".
total unsafe behaviour =
∑ unsafe behaviour
∑ participat ions
Database would
display yesterday
responses for both
questions and
would show the
percentage of
unsafe behavior.
The statistics was
shown in the page
of "statistics".
Yes
Yes
New data was
submitted?
No
End
Figure 2. The flowchart for “1-min Observation” programme
VOLUME Six NUMBER two july - december 2008 PLATFORM
57
Technology Platform: SYSTEM OPTIMISATION
Figure 3. The main page of the “1-min Observation” program
Figure 4. The data sharing page with all employees of Company X
58
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: SYSTEM OPTIMISATION
possible trend in the data collected. At the front page
of the on-line database, only two questions would
be prompted. Two questions per day were adopted
after the consideration of the human factor. The
program was intended to let people feel that it was
so simple to participate and “why not take part?” The
HSE Committee felt that three questions may cause
irritation and one question was just too little and not
efficient in data collection. After much consideration,
two questions were considered the most appropriate.
The questions consisted of areas like Personal
Protection Equipment (PPE) usage, ergonomics, safety
rules, safe practice and housekeeping. The questions
were custom designed to suit Company X interests
to reduce at-risk behaviour of their employees. The
flowchart of the “1-min Observation” program is
shown in the Figure 2. The main page of the “1-min
Observation” program and the data sharing page with
all the employees are given in Figure 3 and Figure 4
respectively.
Participation
The participation from the employees at the early
stage of launching was not good. This was primarily
due to the unfamiliarity of the program and the
routine of going into the web page for the “1-min
Observation” file everyday. After much explanation
and encouragement by the safety manager to
the employees, the response started to increase.
The responses of the employees from respective
departments are shown in Figure 5.
The responses were expected to be low during
weekends and public holidays. Figure 6 shows the
responses received from all the employees and the
responses were on the lower side during weekends.
The trend showed that there was improvement in their
participation over time. However, there were some
occasions when feedback were lower due to some
visitors’ plant visit or corrupted master file. The master
file which was compiled using Microsoft Excel was
easily corrupted due to multiple sharing with many
people and the huge size of the file. The problem was
then fixed by using a standby master file, consistently
backed up with double password protection. Also,
the file size was then reduced by removing some
unnecessary decorative pictures in the file.
Based on the record from the Human Resource
Department, the daily attendance of the employees
was used to compare against the participation rate.
Figure 7 shows the percent of participation relative
to the attendance. The highest participation received
was 86% whereas sometimes it went below 10%. This
depended heavily on the plant activities. If the plant
was experiencing some problems then the response
Figure 5. The responses received from all the respective departments.
VOLUME Six NUMBER two july - december 2008 PLATFORM
59
Technology Platform: SYSTEM OPTIMISATION
Figure 6. The responses received from the employees.
Figure 7. The percentage of participation based on daily attendance.
would be lower as most of the employees were tied
up on the rectification of plant problems.
Data Analysis
The feedback from the employees were analysed
by the database. Each question was prompted four
times. The calculation is shown below.
Total Unsafe Behaviours Observed
Percent of Unsafe Behaviours =
× 100%
Total Responses
60
where
Total Unsafe Behaviours Observed = Unsafe Behaviours Observed Time 1 +
Unsafe Behaviours Observed Time 2 +Unsafe Behaviours Observed Time 3 +
Unsafe Behaviours Observed Time 4
Total Responses = Number of responses Time 1 + Number of responses Time 2
+ Number of responses Time 3 + Number of responses Time 4
From this, the percentages of unsafe behaviours were
sorted accordingly. However, it was noted that even
the topmost unsafe behaviour was only 35% of the
total response as shown in Figure 8. Generally, most of
the respondents were practising the safe behaviour.
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: SYSTEM OPTIMISATION
From Figure 9, the highest unsafe behaviour observed
was the usage of handphones at restricted areas,
which contributed to 35% of the responses. There was
no occasion whereby anyone was observed lifting
goods with improper position.
Discussion
The employees of Company X gave good support and
responses to the “1-min Observation” program. Using
the e-ARBAIS program, it was rather easy to identify
at-risk behaviours that needed improvement. It did
not involve many additional resources to gather the
useful data. As this is a preliminary result, the program
was considered quite successful. A longer time is
needed to see a long–term impact of e-ARBAIS.
Figure 8. The percentage of unsafe behaviours observed based on the total feedback
Figure 9. The top five unsafe behaviours observed
VOLUME Six NUMBER two july - december 2008 PLATFORM
61
Technology Platform: SYSTEM OPTIMISATION
Some challenges in implementation of e-ARBAIS in
Company X and its limitations are shared below.
The Challenges to Implement the e-ARBAIS in an
Organisation
a. Ensure clear communication
As e-ARBAIS was a new concept, it was very
important to communicate clearly to the employees
about the implementation. During the case study,
unclear communication and explanation from safety
department were some of the feedback quoted from
the employees.
b. Ensure management commitment
Management commitment was undeniably an
important role. Management commitment on the
participation would lead others to follow. Consistent
management commitment from each level was
imperative.
c. Ensure follow up action
The e-ARBAIS program could be more effective if
the employees could see the follow-up action by the
Safety Department or HSE Committee. Employees
who participated in the program would be eager
to report their observations and wanted to see the
changes. Thus, if the HSE committee was not able
to take appropriate action to make the changes,
employees would begin to feel disappointed with
the management as there was no follow up action.
Eventually, the program may cease. There was no
motivation that could continue to thrill the employee
to participate in the program.
The Limitation of “1-min Observation” Program
As the case study was rather short, it was difficult to
measure if there was any improvement in the safety
behaviour in the long term.
Also, the “1-min Observation” programme was an IT
based programme. It was thus vital that the database
worked appropriately. During the case study, the
database was created using Microsoft Excel and the
file was corrupted several times and interrupted the
program. Additionally, the file was shared among 79
employees and only one was accessible at one time.
A lot of time was wasted while waiting. Some of
them gave up when they could not open the file on a
particular day. The file malfunctions were occasionally
due to the huge size of the database. Improvement
on this was required to make the programme more
successful.
The questions in the database were developed based
on the recommendations by the company and were
focused on the observation of unsafe act. However,
one of the questions, “Did you experience back pain
after work?” is the result of unsafe acts and might
not be suitable to be included in the database for
the observation purposes. However, it shows the
flexibility of the e-ABAIS that can also be extended to
cover general safety issues resulting from an unsafe
act.
Some of the questions may not be directly understood
by the employees. For instance, “Did you see people
lift things in improper positions?” The employees
might not fully understand what “improper position”
means and therefore inaccurate responses might be
given. The refinement of these questions is required
to ensure accurate feedback from the observers.
d. Honest Participation
The data collected would be very useful if everybody
participated and responded honestly. However, there
was a possibility that people did not answer based on
the observations. Generally, this should not happen
and the overall analysis would be useful.
62
Additionally, there was a possibility that one unsafe
act was observed by multiple observers which
created a false statistic in the database. In this case, a
“feedback” column that enables the observer to write
and describe the observation would definitely help to
minimise the problem.
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: SYSTEM OPTIMISATION
Conclusion
could be more effective if the recommendations were
considered and adopted.
In conclusion, the concept of e-ARBAIS was to serve
as another alternative to the current BBS programme.
It required fewer resources, more involvement, and
low cost which all added on to the sustainability of
the program. The e-ARBAIS was easy to implement
and also in collecting data. It also emphasised and
reminded employees on conducting task with the
correct behaviour. This was intended to provide a
psychological effect to the employees on the safe
behaviour and inculcate the habitual awareness.
Thus, this can encourage a safety culture in an
organization.
The case study of implementing e-ARBAIS in Company
X, which was named as “1-min Observation” had
received positive support and feedback. There were
some constraints in fully implementing the “1-min
Observation” such as, a well designed database,
effective communication between safety department
with the employees and efficient follow up action
on the data analysed. All these would add up to
the success of the case study. It can be further
fine tuned and used in many organisations. Some
recommendations were given below. The programme
Get
ready new
questions
Update
e-ARBAIS
Overall, the e-ARBAIS concept was feasible and
practical. Given a longer time and with the
implementation stage improved, the e-ARBAIS would
definitely benefit the organisation.
Recommendations
a. Appropriate planning for the e-ARBAIS
programme
Most of the employees would like to be informed
about the analysis after their participation. They
wanted to know more about the findings and what
were the unsafe behaviours that were observed most
frequently. Therefore, more proper planning from the
safety department was required. The HSE committee
should take immediate action once the employees
had completed the feedback. The timely analysis
should be shared with everybody. Also, appropriate
action must be taken to show to the employees that
their feedback was valuable. No one would like to
waste their time if they knew nothing was going to
happen with their feedback. It was very important to
note that the sustainability of the program depended
on the confidence level of the employees on the
program.
The flow chart in Figure 10 shows the appropriate
cycle in the e-ARBAIS program.
b. Improvement on database and feedback
column
Share
data &
action
Feedback
Execute
Action
Item
As mentioned earlier, one of the weaknesses during
the case study was the frequent corruptions and
malfunctions of the database.
To overcome this problem, a more stable programme
should be used. A web-based database is much more
user-friendly in this case.
Team
Review
Figure 10. The flow chart of how the e-ARBAIS programme
should be implemented
Additionally, it would be value enhancing if there
were some feedback columns on top of the questions
VOLUME Six NUMBER two july - december 2008 PLATFORM
63
Technology Platform: SYSTEM OPTIMISATION
posted so that the respondents could more accurately
describe the problems and give feedback to the safety
department.
Azmi Mohd Shariff received his MSc in
Process Integration from UMIST, United
Kingdom in 1992. He furthered his studies
at University of Leeds, United Kingdom and
received his PhD in Chemical Engineering
in 1995. He joined Universiti Kebangsaan
Malaysia in 1989 as a tutor upon his return
from Leeds and was appointed as lecturer
in 1996.
c. Sustainability of the program
It is important to ensure that the program is sustainable.
One of the recommendations was to give rewards to
the employees for their feedback given to the program.
Rewards may help to encourage participation and
continuous feedback. In the long term, the safety
program would sustain and the unsafe behaviour of
the employees could be improved.
Acknowledgement
The authors would like to thank Tuan Haji Mashal Ahmad and his
staff on their valuable contribution in this work.
References
[1]
American Psychological Association (APA) (2003).
“Behaviour analyses help people work safer”, Washington.
www.apa.org.
[2]
Geller, E. S. (2002). “How to get people involved in BehaviourBased Safety – selling an effective process”, Cambridge, MA:
Cambridge Center for Behavioural Studies.
[3]
Gilmore, Michael R., Perdue, Sherry R., Wu, Peter (2002).
“Behaviour Based Safety: The next step in injury prevention”.
SPE International on Conference on Health, Safety &
Environment in Oil and Gas Exploration and Production,
Kuala Lumpur, MAL, 20-22 Mar 2002.
[4]
Hayes, S. C. (2000). “The greatest dangers facing behaviour
analysis today”. The Behaviour Analyst Today, Volume 2, Issue
Number 2. Cambridge Center for Behavioural Studies.
[5]
Minshall, S (1997). “An opportunity to get ahead of the
accident curve”. Mine safety and Health News, Vol 4, No. 9.
[6]
Minter, S. G. (2004). “Love the process, hate the name,
Occupational Hazards”, 3rd June 2004.
[7]
SCF Arizona Loss Control (2004). “Behavioural Safety: The
right choice – Listen to your conscience and eliminate
dangerous behaviours”, SCF Arizona, 2nd July 2004, http://
www.scfaz.com/publish/printer_549.shtml
[8]
Schatz, J. R. (2003). “Behaviour-Based Safety: An introductory
look at Behaviour-Based Safety”, Air Mobility Command’s
Magazine, Jan/Feb 2003.
[9]
Skinner, B. F. (1938). “The behaviour of organisms: An
experimental analysis”. Acton, Mass.: Copley Publishing
Group.
64
He joined Universiti Teknologi PETRONAS (UTP) in 1997 and was
appointed as the Head of Industrial Internship in 1998. He was
later appointed as the Head of Chemical Engineering Programme
1999–2003. He is currently an Associate Professor in the
Department of Chemical Engineering and a Leader of the Process
Safety Research Group.
He teaches Process Safety and Loss Prevention at undergraduate
level and Chemical Process Safety at graduate level. He started
research work in the area of Process Safety in 2002. He has
successfully supervised and is currently supervising a few PhD, MSc
and final year undergraduate students in the area of Quantitative
Risk Assessment (QRA), Inherent Safety and Behaviour Based
Safety. He has presented and published more than 15 articles
relating to process safety in conferences and journals. He currently
leads an E-Science research project under Ministry of Science and
Technology entitled ‘Development of Inherent Safety Software for
Process Plant Design’. He has also experience in conducting QRA
for a few local companies in Malaysia. Recently, he was successful
in conducting a short-course on ‘Applied QRA in Process Industry’
to OPU and Non-OPU staff.
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: SYSTEM OPTIMISATION
Bayesian Inversion of Proof Pile Test:
Monte Carlo Simulation Approach
I. S. H. Harahap*, C. W. Wong1
*Universiti Teknologi PETRONAS, 31750 Tronoh, Perak Darul Ridzuan, Malaysia.
1Malaysia LNG Sdn. Bhd., Malaysia.
ABSTRACT
Pile load test is commonly conducted during both design and construction stages. The objective of pile load
test during design stage is to obtain the actual soil parameters and ultimate load in-situ. On the other hand, for
the test conducted during construction stage the objective is to prove that the actual pile capacity conforms to
the design loads. This paper presents probabilistic interpretation of proof pile test to obtain the ultimate pile
capacity. As the first step, the “actual” field parameters are back calculated using ultimate pile capacity from
proof pile load tests. The probabilistic inverse method is used for back calculation of parameters. Soil parameters
obtained from back calculation are then sampled using Monte Carlo simulation technique to generate histogram
of ultimate pile capacity. From the synthetic histogram, other statistical metrics such as mean, standard deviation
and cumulative probability density of ultimate pile capacity can be obtained.
Keywords: Socketed drilled shaft, Monte Carlo simulation, probabilistic inverse analysis, proof pile load test
Introduction
During construction, proof tests to verify pile design are
conducted. Specification usually calls for the amount
of absolute and permanent pile displacement during
a proof test to be less than a specified amount. The
number of piles subjected to proof tests is proportional
to the total number of piles being constructed and
the number of piles that ‘failed’ relative to the number
of proof tests should not exceed a certain prescribed
number. Load applied for proof test is usually twice
the design load at constant loading rate using one or
two load cycles.
obtain actual soil parameters in the form of its joint
probability density. Parameters obtained were then
utilised to generate histograms of ultimate pile load
capacity using Monte Carlo simulation technique.
The arrangement in this paper are as follows: Section
2 outlines geotechnical aspects of the drilled shaft
particularly its design methodology and interpretation
of pile-load-test results at project site near Kuala
Lumpur. The soil condition and pile-load-test results at
project site near Kuala Lumpur are explained in Section
3. In Section 4, the salient features of the probabilistic
inverse method are given, followed by its application
to interpret pile-load-test results in Section 5. Section
6 concludes results from this work.
This paper attempts to interpret pile proof test for
obtaining soil parameters (soil and rock unit skin
resistance as well as base resistance). The probabilistic
inverse analysis method as given in [1] was used to
This paper was presented at the International Conference on Science & Technology: Application in Industry & Education 2008 (ICSTIE 2008), Penang,
12 - 13 December 2008
VOLUME Six NUMBER two july - december 2008 PLATFORM
65
Technology Platform: SYSTEM OPTIMISATION
Geotechnical Aspects
Design of Socketed Drilled Shaft
The ultimate capacity of socketed drilled shaft can be
determined using the following equation:
Q u + Q fu + Qbu
Q u = (f S C S + FRCR) + qbAb
(1)
where Q u is ultimate pile capacity, Q fu is ultimate shaft
capacity, Q bu is ultimate base capacity. The ultimate
skin resistance consist of contribution from soil part
(f S) and rock part (f R) where f S and f R are unit shaft
resistance, C s and CR are circumferential area of pile
embedded in each layer (soil and rock). qb is unit base
resistance for the bearing layer (rock) and Ab pile base
area. Figure 1 shows components of ultimate capacity
of drilled shaft socketed into rock.
The unit skin resistance of cohesionless material
usually has the form of f R = K S σ O tanø where K S is
coefficient of lateral pressure, σ O is vertical overburden
pressure and ø is friction angle [2]. For cohesive
material the unit skin resistance is commonly taken
as proportional to the undrained shear strength as
f R = αS u where α is proportional coefficient and S u
is undrained shear strength [3]. For rock, the unit
skin resistance is empirically determined from rock
unconfined compressive strength, qu [4], or Rock
Quality Designation, RQD. Table 1 shows available
empirical correlation to determine f R from qu and
D
fS
LS
fR
LR
Table 2 to determine f R from RQD. However, the
unit skin resistance has a limiting value depends on
the unconfined compressive strength of the rock
[5]. The empirical correlations proposed is either has
linear relation with qu or power-curve relation to qu .
Evaluation by [6] indicated that the SPT N-value may not
a good indicator of f R due to its sampling rate because
it is too infrequent and suffers too much variability.
From evaluation of pile load test results [7], there is
a significant difference of load-settlement behaviour
among sedimentary, granitic and decomposed rock,
and hence the range of qu for these rocks. Generally
granitic rock has a softer response compared to
sedimentary rock. The ultimate unit resistance range
between 6 to 50 MPa for granitic rock compared to
between 1 to 16 MPa for sedimentary rock. Empirical
correlation by [8] in Table 2 is the lower bound for
sedimentary rock [7]. The rock unit skin resistance, f R
and RQD relationship as in Table 2 is commonly used in
Malaysia to design drilled shaft socketed into rock [9].
Other note, the unit resistance for uplift load should
be adjusted due to contraction of pile under uplift
load, and hence reducing confinement stress [10].
Design approach for socketed drilled shaft varies from
place to place [11]. For example the ultimate capacity
could be determined by considering all resistances
(soil and rock skin resistances, and base resistance) or
totally omitting the base resistance. It is due to the
fact that less displacement is required to mobilise
skin resistance compared to the displacement that
required mobilising base resistance. On the practical
side, the length of rock socket and hence total pile
length is determined in situ during construction based
on observed rock condition, i.e. RQD at that particular
location.
Discussions on various construction method and
constructability issues of drilled shaft can be found in
Ref [12], and the effect of construction method on skin
and base resistance in Ref [13,14]. However, from pull
out test results [6], drilled shaft construction method
has no significant effect on ultimate resistance.
qb
Figure 1. Ultimate capacity of socketed drilled shaft
66
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: SYSTEM OPTIMISATION
Table 1. Empirical value of unit skin resistance for socketed drilled shaft determine from rock unconfined compressive strength and SPT
N-value. Complete references are given in [6]
No.
Empirical Correlation
Reference
1
f R(tsf) = 1.842 qu 0.367
Williams et al. (1980)
2
f R(tsf) = 1.45 qu for clean sockets, and
f R(tsf) = 1.94 qu for rough sockets
Rowe and Armitage (1987)
3
f R(tsf) = 0.67 qu
Horvath and Kenney (1979):
4
f R(tsf) = 0.63 qu
Carter and Kulhawy (1988)
5
f R(tsf) = 0.3 (qu)
Reynolds and Kaderabek (1980):
6
f R(tsf) = 0.2 (qu)
Gupton and Logan (1984)
7
f R(tsf) = 0.15 (qu)
Reese and O’Neill (1987)
8
f R = 0.017V (tsf), or
f R = – 5.54 + 0.4 N(tsf)
Crapps (1986)
9
N-value = {10, 15, 20, 25, 30, >30}
f R(tsf) = {0.36, 0.77, 1.1, 1.8, 2.6, 2.6}
Hobbs and Healy (1979)
10
N-range = 10 - 20, 20 - 50, 50 - 50/3 in., >50/3 in.
f R(tsf) = 1.5, 2.5, 3.8, 5
McMahan (1988)
Pile Load Test
Pile-load-test generally serves two purposes. When
conducted during design stage, it helps to establish
parameters to be used in design, and when conducted
during construction stage, it proves working
assumptions during design. To obtain the “actual
soil parameters” for design purpose, the pile test is
instrumented [4] and parameters are back calculated
from data obtained during testing. The number of
this type of test is limited due to its cost; therefore
site variability of pile ultimate capacity cannot be
established.
Table 2. Empirical value of unit skin resistance for socketed drilled
shaft determine from RQD Ratio
RQD Ratio %
Working Rock Socket Resistance
f R (kPa)
Below 25
300
25 - 70
600
Above 70
1000
The load settlement curves do not always show a sign
of failure as specified by various methods, for example
Davisson’s, Terzaghi’s, Chin’s methods and others. As
such, it is difficult to ascertain the validity of design
assumption, i.e. the ultimate skin resistance, based on
information’s obtained from this test. Other elaborate
method to interpret pile load test in sand and clay
can be found in [15, 16]. A new and novel approach
that utilizes a data base of pile load tests is recently
proposed by [17]. In their method design parameters
are extracted from the data base using Bayesian neural
network that intelligently update its knowledge when
new information is added to the data base.
Other than instrumented pile as previously cited, [18]
proposed method to derive soil parameters from load
settlement curve. The approach uses “projected load
settlement curve” to obtain the ultimate capacity.
The projected load settlement curve is an analytical
function for load settlement relations with parameters
of the function are obtained from regression of the
actual load settlement curve. For this purpose, failure
VOLUME Six NUMBER two july - december 2008 PLATFORM
67
Technology Platform: SYSTEM OPTIMISATION
is defined as correspond to a settlement of 10% of pile
diameter.
As a proof test, pile-load-test results are interpreted
based on criteria establishes to achieve the
design objectives and elucidated in the technical
specification. As an example, the settlement of the
pile tested should not exceed specified settlement at
working load and permanent settlement should not
exceed settlement at twice of the working load. The
pile is considered “pass” if both criteria are satisfied.
The number of proof pile test is prescribed based on
total length of pile constructed; as such more than one
proof pile-load-test is conducted within one project.
Furthermore, it is common that proof pile-load-test is
“fail to reach failure”, in other words the applied load
is less than ultimate capacity of the pile.
While the interpretation of proof test based on
settlement criteria lay out in the technical specification
is sufficient for practical purposes, there are also
attempts to further exploit information’s from proof
pile-load-test. For example, from pile-load-test that
reaches failure, information’s can be obtained to
update the reliability of pile [19-22]. For pile-loadtest that fails to reach failure, information’s can be
obtained to update the probability distribution of
pile capacity [23]. These approaches follow trend
of migration of geotechnical analysis from factor of
safety based to reliability based [24]. It is worth to
note that, besides ultimate load limit state approach
previously cited, serviceability limit state for drilled
shaft to establish probability of failure and reliability
index from load settlement curve of pile load test also
have been attempted in [25,26]. In their approach, the
pile load settlement curves are calculated using “t-z”
approach and finite difference method. Probabilistic
load-settlement curves are developed using Monte
Carlo simulation. From the histogram generated,
the probability of failure and reliability index can be
determined.
For the work reported herein, the probability density
of soil parameters are calculated from ultimate
pile capacity, deduced from pile-load-test, using
probabilistic inverse method. Monte Carlo simulation
technique is then used to generate the histogram
of pile capacity. The probability of ultimate pile
capacity, or the reliability index, can be obtained from
cumulative probability density of pile capacity.
Description of the Project
Site Condition
A total of 12 bore holes were carried out for
soil investigation during design stage. The soil
investigations were mainly carried out by using
Standard Penetration Test (SPT) as well as standard soil
index and physical properties tests. The site condition
was mainly formed by 3 types of soil, which were silt,
clay and sand. Silt was found on the top of the soil layer
Figure 2. Load settlement curve for proof pile-load-test. Davisson’s ultimate load can be obtained only in two out of nineteen tests.
68
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: SYSTEM OPTIMISATION
while very stiff or hard sandy silt were encountered
on the next layer which range from Reference Level
(RL) 70 m to RL 50 m. Generally, high organic content
was observed for the upper layer materials. The RQD
for the site ranged from 10% to 30%, with average of
25%.
Pile Loading Test
There were three types of pile being used at the site
consisting of a 450 mm, 600 mm and 900 mm diameter
bored pile with design load of 1500 kN, 4000 kN and
9000 kN respectively. Out of a total of 19 proof pileload-tests conducted, only two gave ultimate pile
capacity based on Davisson’s criteria. The calculated
ultimate pile capacity is 8900 kN and 3900 kN for 900
mm and 600 mm diameter piles, respectively. Figure
2 shows load deflection curves and ultimate load
determination procedure using Davisson’s method
for 600 m and 900 mm piles, and Table 3 shows the
schedule of all tests. It should be noted that for both
the 600 mm and 900 mm piles one out of seven test
piles being tested was fail.
Probabilistic Inverse Method
Suppose that we have function f that map parameters
into theoretical quantity such that d = f(m) where d
= {di ,…, d ND and m = {mi , ,…, m NM}, the objective
of inverse analysis is to determine m given m. In the
Table 3. Attributes of pile proof test result
Estimate of Q u
(kN)
Pile Length
(m)
FF1- P28
Not fail
6.000
900
4.5
FF1- P67
8900
10.245
900
4.5
FF2 - P184
Not fail
6.000
900
4.5
FF2 - P472
Not fail
11.000
900
4.5
FF3 - P234
Not fail
11.075
900
4.5
FF4 - P236
Not fail
12.325
450
1.5
FF4 - P275
Not fail
14.625
450
1.5
FF5 - P33
Not fail
12.425
600
3.0
FF5 - P85
Not fail
22.125
600
3.0
FSK 1,5 - P47
Not fail
14.000
900
4.5
FSK 1,5 - P44
Not fail
8.200
450
1.5
FSK 2,3,4 - P200
Not fail
8.800
900
4.5
FSK 2,3,4 -P387
Not fail
18.000
900
4.5
FSK 6 - P 74
Not fail
2.600
900
2.6
FSK 6 - P 307
Not fail
5.700
600
3.0
FSK 7 - P40
Not fail
21.025
600
3.0
FSK 7 - P370
Not fail
7.330
600
3.0
PSB - P138
Not fail
19.500
600
3.0
PSB- P183
3900
10.775
600
3.0
Pile Location
Pile Diameter
(mm)
Socket Length
(m)
VOLUME Six NUMBER two july - december 2008 PLATFORM
69
Technology Platform: SYSTEM OPTIMISATION
context of pile-load-test, to determine f S , f R and qbu
knowing Q u obtained from pile-load-test and f is the
relationship in Eq. (1).
Data Space
Suppose that we have observed data values dobs, the
probability density model to describe experimental
uncertainty, such as Gaussian model, can be written
as follows:
 1

(d d obs )T C D1 (d d obs )   2

where CD is the covariance matrix. If the uncertainties
are uncorrelated and follow Gaussian distribution, it
can be written as
2
 d i d i  

obs 

 
i

 
(3)
Model Space
In a typical problem we have model parameters that
have a complex probability distribution over the
model space. The probability density is denoted as
ρM(m). Suppose that we know joint probability density
function ρ(m,d) and d = f(m), then the conditional
probability density function,
σM(m) = ρM|d(m)(m|d = f(m))
can be obtained as follows [1]:
M (m )
= k (m , f (m ))
det ( g M
F Tg D F)
det g M det g D
g M F Tg D F
1/ 2
gM .
gD
1/ 2
.
d =f (m )
For constant gM(m) and gD(d), and linear or weak
linearity problem [1], Eq. (4) reduces to
Other approach is to use Markov Chain Monte Carlo
(MCMC) simulation that generate sampling points
over the model space by “controlled random walk”,
the Markov Chain, that eventually converged to the
conditional probability distribution of the posterior (or
parameters). In Markov Chain approach the sequence
of random variables X0, X1, X2, … at each time
t ≥ 0the next state Xt+1 is sampled from a distribution
P(Xt+1 | Xt) that depends on the state at time t. Similar
to Monte Carlo simulation, if sufficient numbers of
sampling points are obtained, then the approximation
to the expected value is evaluated through Eq. (6).
In general the MCMC has three basic rules:
(i) a proposal Markov Chain rule expressed by a
transition kernel q(x,y),
(5)
(ii) an accept reject rule which accept or reject a
newly proposed Y k = q(X k ,.)where X k is recently
accepted random variable, and
(iii) a stopping rule.
A typical MCMC has the basic algorithm shown as the
following algorithm
where k is the normalizing factor, µD(d) is homogenous
probability density function, which upon integration
over the data space become unity.
70
(6)
1/ 2
(4)
D (d )
M (m ) = k M (m )
D ( d ) d =f (m )
The analytical form of posterior distribution, i.e. Eq. (5),
is difficult to obtain, or, even if obtainable, is difficult
to interpret. One has to resort to simulation approach
such as Monte Carlo simulation to obtain parameter
pairs over the model space and used such data for any
application. In Monte Carlo simulation, after sufficient
number on sampling of random variables X0, X1, …, Xn
the expectation µ = E{g(Xi} is approximated as:
(2)
D ( d ) = k exp 

 1
D (d ) = k exp 
 2

Evaluation of posterior distribution
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: SYSTEM OPTIMISATION
Basic MCMC Algorithm
1.
Draw initial state X0 from some initial
distribution
2.
For I = 0 to N do
3.
Modifying Xt according to some proposal
distribution to obtain a proposed sample Yt+1
where Yt+1 = q (Xt , Y)
4.
With some probability A(Xt , Yt+1) accept Yt+1
The effects of prior knowledge on ultimate pile
capacity can be investigated using various forms of
density distribution. In this work, only the effect of
unit skin and base resistances of rock are considered,
and the joint probability density is obtained from Eq
(5) as
Y
with probabilit y A ( X t , Y t +1 )
X t +1 =  t +1
otherwise X t

Some acceptance rules are given as follow
(a) Metropolis sampling:
( Y t +1 )
A ( X t , Y t +1 ) =
M ( fR , q b )
(X t )
(b) Metropolis-Hasting sampling
A ( X t , Y t +1 ) =
( Y t +1 ) + q( Y t +1, X t )
( X t ) q( X t , Y t +1 )
(c) Boltzman sampling:
A ( X t , Y t +1 ) =
probability density model to describe experimental
uncertainty (Eq. 3) is formed using the theoretical
model d = f(m) as in Eq. (1), and observed pile ultimate
capacity as dobs. The joint probability density is then
σM(m) = σM(f S , f R , qb). Prior knowledge can be
incorporated in ρM(m) = ρM(f S , f R , qb) particularly
knowledge on those parameters specific for the rock
type and its locality.
( Y t +1 )
( X t ) + ( Y t +1 )
BAYESIAN Interpretation of proof Pile Load
Test
The simplistic model of bearing capacity of socketed
drilled shaft is given by Eq. (1). Assuming known pile
geometry, the model space is then m = (f z , f R , qb). The
probability density model to describe experimental
uncertainty (Eq. 3) is formed using the theoretical
model d = f(m) as in Eq. (1), and observed pile ultimate
capacity as dobs. The joint probability density is then
σM(m) = σM(f S , f R , qb). Prior knowledge can be
incorporated in ρM(m) = ρM(f S , FR , qb) particularly
knowledge on those parameters specific for the rock
type and its locality.
The simplistic model of bearing capacity of socketed
drilled shaft is given by Eq. (1). Assuming known pile
geometry, the model space is then m = (f S , f R , qb). The
=
M ( fS , fR , q b ) dfS (7)
In this work two aspects are being investigated: (a)
the effect for prior knowledge on predicted pile
bearing capacity and (b) comparison between the
“brute force” Monte Carlo and Markov Chain Monte
Carlo simulations. For the first objective, four cases
are considered. In Cases 1 to 3 lognormal prior
distributions of f R are assumed with mean values range
between 300 to 800 kPa. For the fourth case, normal
distribution is assumed for f R with mean value of 300
kPa. These values conform to an empirical value for
unit skin resistance for rock at low RQD (Table 2). This
number is somewhat lower than back calculated from
pile load test in limestone which range from 900 kPa
to 2 300 kPa [29]
Results from Case 1 to 4 are compared in term of:
(a) plot of posterior probability density, (b) Monte
Carlo sampling points and (c) histogram of predicted
ultimate pile capacity and shown in Figure 3 to 6.
Figure 7 shows relative comparison of prior density
distribution used in Case 1 to 3. A sample of 3D plot is
shown in Figure 8 for Case 1. In Case 5 MCMC is used
to draw sampling points with prior density distribution
as used in Case 4. All cases use 15 000 trials. Results for
all cases are shown in Table 2 for a 600 mm diameter
pile.
VOLUME Six NUMBER two july - december 2008 PLATFORM
71
Technology Platform: SYSTEM OPTIMISATION
(a)
(b)
(c)
Figure 3. (a) Plot of posterior probability density for Case 1, (b) Sampling points superimposed to probability density plot and (c)
Histogram of ultimate pile capacity
(a)
(b)
(c)
Figure 4. (a) Plot of posterior probability density for Case 2, (b) Sampling points superimposed to probability density plot and (c)
Histogram of ultimate pile capacity
(a)
(b)
(c)
Figure 5. (a) Plot of posterior probability density for Case 3, (b) Sampling points superimposed to probability density plot and (c)
Histogram of ultimate pile capacity
72
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: SYSTEM OPTIMISATION
(a)
(b)
(c)
Figure 6. (a) Plot of posterior probability density for Case4, (b) Sampling points superimposed to probability density plot and (c)
Histogram of ultimate pile capacity
Figure 7. Prior distribution of unit skin resistance of rock
Figure 7. A 3D plot of posterior joint probability distribution for
Case 1
Figure 9. Case 5: (a) Plot sampling points generated by Markov chain and (b) Histogram of ultimate pile capacity
VOLUME Six NUMBER two july - december 2008 PLATFORM
73
Technology Platform: SYSTEM OPTIMISATION
Table 4. Comparison of mean, median and standard deviation
Standard
Deviation
(kN)
Remark
Case
Mean
(kN)
Median
(kN)
Case 1
4103
4147
521
Lognormal Distribution, MC
Case 2
4108
4164
586
Lognormal Distribution, MC
Case 3
4125
4189
573
Lognormal Distribution, MC
Case 4
4145
4169
590
Normal Distribution, MC
3840
3905
351
(a) Normal Distribution, MCMC (burned point 100)
3902
3931
270
(b) Normal Distribution, MCMC (burned point 500)
Case 5
Discussions
Concluding Remark
From Table 4 and Figures 3, 4, 5, 6 and 9 the following
observations can be advanced:
Based on previous discussions the following
conclusions can be put forward:
• Method to interpret pile test to obtain
probabilistic characteristics of ultimate load has
been presented. The first step is to obtain joint
probability distribution of soil parameters in the
material (or parameter) space and the second
step is to generate histogram of ultimate pile
capacity using Monte Carlo technique. Statistical
characteristics of the ultimate pile capacity are
then obtained from the synthetic histogram using
standard discrete method.
• From this study the effect of prior probability
distribution, for all practical purposes, is
negligible.
• Markov Chain Monte Carlo method yields more
accurate results compare to brute force Monte
Carlo method and more efficient in term of ratio
of generated sampling points to number of trial.
For all practical purposes the prior density distributions
have minimal effect on calculated ultimate pile
capacity. The mean values ranged between 4 103 to
4 145 kN.
Markov Chain Monte Carlo is more efficient compared
to the brute force Monte Carlo method. Out of 15 000
trials, 6 042 points have been generated using MCMC
method compare to 2 066 points using brute force
Monte Carlo method. The sampling points for MCMC
method are more concentrated around the maximum
probability density (Figure 9a) compare to MC method
(Figure 6b) resulting in a more accurate ultimate pile
capacity.
Markov Chain Monte Carlo yield more accurate results
(3 902 kN to 3 900 kN from pile test) compare to brute
force Monte Carlo method (4 145 kN to 3 900 kN from
pile test).
74
References
[1]
Mosegaard, K. & Tarantola, A. “Probabilistic Approach to
Inverse Problem”. In International Handbook of Earthquake &
Engineering Seismology (Part A), Academic Press. 2002. pp.
237-265
[2]
Rollins, K. M., Clayton, R. J., Mikesell, R. C. & Blaise, B. C.
“Drilled Shaft Side Friction in Gravely Soils”. Journal of
Geotechnical and Geoenvironmental Engineering, 2005.
131(8):987-1003
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: SYSTEM OPTIMISATION
[3]
O’Neil, M. W. “Side Resistance in Piles and Drilled Shafts”.
In The Thirty-Fourth Karl Terzaghi Lecture. Journal of
Geotechnical and Geoenvironmental Engineering, 2001.
127(1):1-16
[17] Goh, A. T. C., Kulhawy, F. H. & Chua, C. G. “Bayesian Neural
Network Analysis of Undrained Side Resistance of Drilled
Shafts”. Journal of Geotechnical and Geoenvironmental
Engineering, 2005. 131(1):84-93
[4]
Zhang, L. & Einstein, H. H. “End Bearing Capacity of
Drilled Shafts in Rock”. Journal of Geotechnical and
Geoenvironmental Engineering. 1998. 124(7):574-584
[18] Boufia, A. “Load-settlement Behaviour of Socketed Piles
in Sandstone”. Geotechnical and Geological Engineering.
2003. 21:389-398
[5]
Amir, J. M. “Design of Socketed Drilled Shafts in Limestone,
a Discussion”, Journal of Geotechnical Engineering. 1994.
120(2):460-461
[19] Kay, J. N. “Safety Factor Evaluation for Single Piles in Sand”.
Journal of Geotechnical Engineering Division. 1976.
102(10):1093-1108
[6]
McVay, M. C., Townsend, F. C. & Williams, R. C. “Design of
socketed drilled shafts in limestone”. Journal of Geotechnical
Engineering. 1992. 118(10): 1626-1637
[7]
Ng, C. W. W, Yaw, T. L. Y, Li, J. H. M. & Tang, W. H. “Side Resistance
of Large Diameter Bored Piles Socketed into Decomposed
Rocks”. Journal of Geotechnical and Geoenvironmental
Engineering. 2001.127(8):642-657
[20] Lacasse, S. & Goulois. A. “Uncertainty in API Parameters
for Predictions of Axial Capacity of Driven Piles in Sand”.
Proceeding of the 21st Offshore Technology Conference,
Society of Petroleum Engineers, Richardson, Texas. 1989.
353-358
[8]
Horvath, R. G. & Kenney, T. C. “Shaft Resistance of Rocksocketed Drilled Piers”. Proceedings Symposium on Deep
Foundation. 1979
[9]
Tan, Y. C. & Chow, C. M. “Design and Construction of Bore Pile
Foundation”. Geotechnical Course for Foundation Design &
Construction. 2003
[10] Fellenius, B. H. Discussion of ‘‘Side Resistance in Piles
and Drilled Shafts’. Journal of Geotechnical and
Geoenvironmental Engineering. 2001. 127(1): 3–16
[11] Hejleh, N. A., O’Neill, M. W, Hanneman, D. & Atwooll,
W. J. “Improvement of the Geotechnical Axial Design
Methodology for Colorado’s Drilled Shafts Socketed in Weak
Rocks”. Colorado Department of Transportation. 2004.
[12] Turner, J. P. “Constructability for Drilled Shafts”. Journal
of Construction Engineering and Management. 1992.
118(1):77-93
[13] Majano, R. E., O’Neill, M. W. & Hassan, K. M. “Perimeter Load
Transfer in Model Drilled Shafts Formed Under Slurry”.
Journal of Geotechnical Engineering. 1994. 120(12.):21362154
[14] Chang, M.F. & Zhu, H. Construction Effect on Load
Transfer along Bored Pile. Journal of Geotechnical and
Geoenvironmental Engineering. 2004. 130(4):426-437
[21] Baecher, G. R. & Rackwitz, R. “Factor of Safety and Pile Load
Tests”. International Journal of Numerical and Analytical
Methods in Geomechanics. 1982. 6(4):409-424
[22] Zhang, L. M. & Tang, W. H. “Use of Load Tests for Reducing Pile
Length”. Proceeding of the International Deep Foundations
Congress. Geotechnical Special Publication No. 116, M. W.
O’Neill and F. C. Townsend, eds., ASCE, Reston, Va., 2002.
993–1005
[23] Zhang, L. M. “Reliability Using Proof Pile Load Tests”. Journal
of Geotechnical and Geoenvironmental Engineering. 2004.
130(2): 1203-1213
[24] Duncan, J. M. “Factors of Safety and Reliability in
Geotechnical Engineering”. Journal of Geotechnical and
Geoenvironmental Engineering. 2000. 126(4):307-316
[25] Misra, A. & Roberts, L. A. “Axial Service Limit State Analysis of
Drilled Shafts using Probabilistic Approach”. Geotechnical
and Geological Engineering, 2006. 24:1561–1580
[26] Misra, A., Roberts, L. A. & Levorson, S. M “Reliability Analysis
of Drilled Shaft Behaviour Using Finite Difference Method
and Monte Carlo Simulation”. Geotechnical and Geological
Engineering. 2007. 25:65–77
[27] Hassan, K. M. & O’Neill, M. W. “Side Load-Transfer Mechanisms
in Drilled Shafts in Soft Argillaceous Rock”, Journal of
Geotechnical and Geoenvironmental Engineering. 1997.
123(2):145-152
[15] Cherubini, C., Giasi, C.I. & Lupo, M. Interpretation of Load
Tests on Bored Piles in the City of Matera. Geotechnical and
Geological Engineering. 2004. 23:239-264
[28] Hassan, K. M., O’Neill, M. W., Sheikh, S. A. & Ealy, C. D. “Design
Method for Drilled Shafts in Soft Argillaceous Rock”. Journal
of Geotechnical and Geoenvironmental Engineering. 1997.
123(3):272-280
[16] Pizzi, J.F. Case history: Capacity of a Drilled Shaft in
the Atlantic Coastal Plain. Journal of Geotechnical and
Geoenvironmental Engineering, 2007. 133(5):522-530.
[29] Gunnink, B. & Kiehne, C. “Capacity of Drilled Shafts in
Burlington Limestone. Journal of Geotechnical and
Geoenvironmental Engineering”. 2002. 128(7):539-545
VOLUME Six NUMBER two july - december 2008 PLATFORM
75
Technology Platform: SYSTEM OPTIMISATION
I. S. H. Harahap holds a Bachelor’s Degree
(Sarjana Muda, SM) and Professional
Degree (Insinyur, Ir.) from Universitas
Sumatera Utara, Medan, Indonesia. After a
short stint with the industry, he continued
his tertiary education in the United States,
obtained his Master’s of Science in Civil
Engineering (MSCE) degree from Ohio
University, Athens, Ohio and Doctor of
Philosophy (PhD) degree from Northwestern University, Evanston,
Illinois. He was with Universitas Sumatera Utara (USU) before
joining Universiti Teknologi PETRONAS (UTP) in August 2005.
His research interests include: (1) application of expert system in
geotechnical engineering, (2) implementation of geotechnical
observational method, (3) subsurface exploration and site
characterization, (4) landslide hazard identification and mitigation,
and (5) robust and reliability based structural optimization. He
is the recipient of 1993 Thomas A. Middlebrook Award from
American Society of Civil Engineer (ASCE) for his research on
simulation of the performances of braced excavation in soft clay.
76
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Wong Chun Wah graduated with a
Bachelor of Science degree in Civil
Engineering from Universiti Teknologi
PETRONAS (2008). Currently, he is working
with PETRONAS’s Malaysia LNG Sdn. Bhd. in
Bintulu as a Civil and Structural Engineer.
Technology Platform: SYSTEM OPTIMISATION
ELEMENT OPTIMISATION TECHNIQUES
IN MULTIPLE DB BRIDGE PROJECTS
Narayanan Sambu Potty*, C. T. Ramanathan1,
*Universiti Teknologi PETRONAS, 31750 Tronoh, Perak Darul Ridzuan, Malaysia
1Kumpulan Liziz Sdn Bhd, Kota Kinabalu, Sabah, Malaysia.
*[email protected]
Abstract
Management of multiple bridge projects relies on the selection and optimisation of elements. Problems are
avoided if construction knowledge and experience are utilised. A systematic selection process is required
during design stage to maximise the use of available resources. Multiple bridge designs incorporate practical
suggestions from the field personals. The case study in East Malaysia considers problems of material, labour and
equipment availability. The average ratio presence of each element is calculated to show its impact. The need
of element optimisation techniques in bridges is emphasised. The database is presented and also the process
undertaken during the design is discussed.
Keywords: Design and Build process, Multiple project management, Bridge elements, Average ratio presence.
INTRODUCTION
Bridge structures are designed with high quality and
safety standards but sometimes with not enough
attention to construction methods, site conditions and
details. Construction problems encountered during
execution are complex and costly. Many construction
problems can be avoided with proper attention
and consideration of the construction process
during the design phase [1]. Factors of simplicity,
flexibility, sequencing, substitutions and labour skill
and availability should be the part of design. The
appropriate use of standardisation can have several
benefits [1]. These include increased productivity
and quality from the realization of repetitive field
operations, reduction in design time, savings from
volume discounts in purchasing, and simplified
materials management. This method of standardising
bridge elements may be suitable for selective projects
of same nature but are less significant and more
complex when constructing multiple bridge projects
situated at different site conditions.
The construction process and success in management
of multiple bridge projects directly relies on the
selection and optimisation of their elements/
components. A systematic optimization process
is adopted during the conceptual design stage
to overcome the resource constraints during the
construction phase. The knowledge of construction
experience is also utilised during the element
optimisation process.
D&B combines the design and construction functions
to vests its responsibility with one entity: the designbuilder. The D&B process changes some fundamental
relationships between the client and the contractor.
The client employs a Project Director as the
This paper was presented at the International Conference on Construction and Building Technology 2008, Kuala Lumpur,
16 - 20 June 2008
VOLUME Six NUMBER two july - december 2008 PLATFORM
77
Technology Platform: SYSTEM OPTIMISATION
representative, whereas the contractor has to engage
design consultant and manage the construction of the
whole project. Owners employ only one contractor
who is solely responsibility for delivering the assigned
project with defined requirements, standards and
conditions. Both parties are expected to aim for a
successful project outcome.
BACK GROUND AND PROBLEM STATEMENT
Sabah situated on the northern tip of the island of
Borneo is the second largest state in Malaysia. Over
70% of its population lives in rural area as majority are
depending directly or indirectly on agriculture. The
state has several development projects procured by
D&B Method for Upgrading rural roads and Bridges
replacement for the benefit of rural sectors contributing
to the national economy. Five contract packages
comprising 45 bridges located in 12 districts in the
state having 76 115 square kilometer coverage area [2]
and two road projects in one district was assigned to
the first author’s company. As summarised in Figure
1, the Bridge projects and the two road projects were
handled simultaneously.
This study examines the use of element optimisation
technique through a case study for managing the
above multiple DB bridge projects in Sabah, East
Malaysia. The data of bridge elements of all the 45
bridges were collected and compiled. The ratio of
each element in individual bridges and their average
ratio presence in each projects were compared for
study.
RESEARCH METHODOLOGY
The element ratio comparison and analysis were made
in the following steps.
1. Review of all the five projects and compilation of
the summary.
2. Prepare the element ratio table and Pi chart for
each bridge of all the projects and analyse the
ratio of their impact on the individual bridges.
3. Compress and derive a single and common
average ratio table and pi chart showing the
overall impact of each elements for the entire
multiple project.
4. Identify the critical and crucial elements that need
attention.
5. Discussion on the element of major contributions
which is to analyse the element with maximum
element ratio.
ANALYSIS OF COMPONENTS OF MULTIPLE DB
BRIDGE PROJECTS
Schedules of the multiple projects
The projects started at different times having
simultaneous construction periods at various stages
Figure 1. Duration of Multiple D&B projects in Sabah
78
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: SYSTEM OPTIMISATION
Table 1. Schedule of Multiple contracts handled in Sabah
Project
Number
No. of
Bridges
No. of districts
Locations
Months
Period
1
12
3
18
Jul 03 – Jan 05
2
5
3
18
Jan 05 – Jul 06
3
8
3
18
Jul 05 – Jan 07
4
13
3
20
Oct 05– Jun 07
5
7
3
20
Oct 05– Jun 07
7
45
12
48
Jul 03–Jun 07
Total
of completion as shown in Table 1. The five projects
involving 45 Bridges have been completed successfully
and delivered to client for public usage.
Bridge element ratio
Project Duration
Table 2. Weight of each Element in bridges of Package A
Description
Ratio
Foundation (Piling)
22.64%
Abutment and Wing Wall
16.33%
Piers, Crossheads and Pilecaps
The element ratio for bridges for each project was
calculated as given in the following section.
1.57%
Precast, Prestressed Beam and Ancillary
39.88%
Diaphragms
2.87%
Bridge Deck and Run-On-Slab
13.93%
Parapet
Project 1 – Package A
2.77%
TOTAL
Table 2 and Figure 2 show the percentage presence of
each element in constructing the bridges in Package
A. From the weight breakdown it is clear that critical
elements “Piling work (Foundation)” has greater than
20% presence and production of beams and erection
has nearly 40% presence. The influence of these
elements in the project is high in spite of their low
quantum in physical work (20% and 40% respectively)
because of their specialty nature. Hence, care should
be taken while choosing these specialties works to
suite the local availability and eventually use few
specialised designs for those particular elements. In
this manner these crucial elements in Package A were
optimised in the design as follows:
Critical element No. 1: Piling works
– Steel H piles of driven type for 6 Bridges
– Micro pile of drilling type for 3 Bridges
Critical element No. 2: Beams
– Cast in situ Post tensioned beams of two
shapes I-16 and I-20.
100.00%
3%
13 .9
%
2. 77
4%
22 .6
7%
2. 8
3%
16 .3
8%
39 .8
%
1. 57
Figure 2. Weight of elements expressed as a percentage
VOLUME Six NUMBER two july - december 2008 PLATFORM
79
Technology Platform: SYSTEM OPTIMISATION
Project 2 – Package B
Table 3 and Figure 3 show the percentage presence of
each element in constructing the bridges in Package
B. From the weight breakdown it is clear that critical
elements are “Piling work (Foundation)” which has
greater than 20% presence and “production of beams
and erection” with 30% presence. The reasons as
explained above for Project 1 have eventually resulted
due to the use of specialised designs for those
particular elements. In this manner these crucial
elements in Package B were optimised in the design
as follows:
Critical element No. 1: Piling works
– Micro pile for 3 Bridges
– Spun pile for 1 Bridges
Critical element No. 2: Beams
– Cast in situ Post-tensioned beams of two
shapes I-16 and I-20 for 2 Bridges
– Prestressed precast beams (in factory) for 3
Bridges.
Project 3 – Package C
Table 4 and Figure 4 show the percentage presence of
each element in constructing the bridges in Package
C. From the weight breakdown it is clear that critical
elements are again “Piling work (Foundation)” with
greater than 20% presence and the “production of
beams and erection” with greater than 30% presence.
The reasons as explained above has eventually
resulted due to the use of specialised designs for
those particular elements. In this manner these crucial
elements in Package C were optimised in the design
as follows:
Table 3. Weight of each element in bridges of
Package B
Description
Ratio
Foundation (Piling)
29.41%
Abutment and Wing Wall
12.68%
Piers, Crossheads and Pilecaps
Precast, Prestressed Beam and Ancillary
80
32.50%
Diaphragms
1.63%
Bridge Deck and Run-On-Slab
11.88%
Parapet
3.11%
TOTAL
100.00%
3.11%
11.88%
29.41%
1.63%
32.50%
12.68%
8.79%
Figure 3. Weight of elements expressed in percentage
Table 4. Weight of each element in bridges of Package C
Description
Ratio
Foundation
22.44%
Abutment and Wing Wall
12.13%
Piers, Crossheads and Pilecaps
13.56%
Precast, Prestressed Beam and Ancillary
31.73%
Diaphragms
2.64%
Bridge Deck and Run-On-Slab
13.84%
Parapet
3.66%
TOTAL
100.00%
3.66%
13.84%
Critical element No. 1: Piling works
– Bored pile for 3 Bridges
– Micro pile for 2 Bridges
– Spun pile for 1 Bridges
Critical element No. 2: Beams
– Cast in situ Post tensioned beams of two
shapes I-16 and I-20 for 4 Bridges.
– Prestressed precast beams (in factory) for two
bridges.
8.79%
22.44%
2.64%
12.13%
31.73%
13.56%
Figure 4. Weight of each element expressed
in percentage
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: SYSTEM OPTIMISATION
Project 4 – Package D
Table 5. Weight of each element in bridges of Package D
Table 5 and Figure 5 show the percentage presence of
each element in constructing the bridges in Package
D.
Critical element No. 1: Piling works
– Micro pile for 9 Bridges
– Spun pile for 2 Bridges
Critical element No. 2: Beams
– Cast in situ Post tensioned beams of two
shapes I-16 and I-20 for 6 Bridges.
– Prestressed precast beams (in factory) for 6
bridges.
From the weight breakdown of the elements of
bridge, it is clear that critical elements are “Piling
work (Foundation)” has greater than 30% presence
and the “production of beams and erection” with
30% presence. The reasons as explained above has
eventually resulted due to the use of specialised
designs for those particular elements. In this manner
these crucial elements in Package D were optimised in
the design as follows:
Description
Ratio
Foundation (Piling)
33.82%
Abutment and Wing Wall
16.24%
Piers, Crossheads and Pilecaps
1.42%
Precast, Prestressed Beam and Ancillary
31.36%
Diaphragms
2.27%
Bridge Deck and Run-On-Slab
12.15%
Parapet
2.73%
TOTAL
100.00%
2.27%
12.15%
2.73%
31.36%
33.82%
1.42%
16.24%
Figure 5. Weight of each element expressed in percentage
Table 6. Weight of each element in bridges of Package E
Description
Ratio
Project 5 – Package E
Foundation (Piling)
21.05%
Abutment and Wing Wall
8.67%
Table 6 and Figure 6 show the percentage presence
of each element in the construction of the bridges in
Package D. From the weight breakdown it is clear that
critical elements are again “Piling work (Foundation)”
with greater than 30% presence and the “production
of beams and erection” with presence of greater than
55%. The reason is due to the design using Steel girder
to suite the site condition. In this manner these crucial
elements in Package E were optimised in the design
as follows:
Piers, Crossheads and Pilecaps
1.78%
Critical element No. 1: Piling works
– Micro pile for 5 Bridges
– Spun pile for 1 Bridge
Critical element No. 2: Beams
– Cast in situ Post tensioned beams I-16 for 1
Bridge.
– Steel girders 5 Bridges
– Steel trusses 1 Bridge
Precast, Prestressed Beam and Ancillary
56.84%
Diaphragms
2.16%
Bridge Deck and Run-On-Slab
7.77%
Parapet
1.72%
TOTAL
100.00%
7.77%
2.16%
1.72%
21.05%
8.67%
1.78%
56.84%
Figure 6. Weight of each element expressed as a percentage of
total
VOLUME Six NUMBER two july - december 2008 PLATFORM
81
Technology Platform: SYSTEM OPTIMISATION
Table 7. Overall weight of each element for bridges in Package A to E
Ratio
Description
Package E
Average
33.82%
21.05%
25.87%
12.13%
16.24%
8.67%
13.21%
13.56%
1.42%
1.78%
5.43%
31.73%
31.36%
56.84%
38.46%
1.63%
2.64%
2.27%
2.16%
2.32%
13.93%
11.88%
13.84%
12.15%
7.77%
11.91%
2.77%
3.11%
3.66%
2.73%
1.72%
2.80%
100.00%
100.00%
100.00%
100.00%
100.00%
100.00%
Package A
Package B
Package C
Foundation (Piling)
22.64%
29.41%
22.44%
Abutment and Wing Wall
16.33%
12.68%
1.57%
8.79%
39.88%
32.50%
2.87%
Piers, Crossheads and Pilecaps
Precast, Prestressed Beam and Ancillary
Diaphragms
Bridge Deck and Run-On-Slab
Parapet
TOTAL
Discussion of Results
2.80%
11.91%
25.87%
2.32%
13.21%
38.46%
5.43%
Figure 7. Overall Weight of each element expressed in
percentage
Overall average ratio
The Table 7 and Figure 7 show the overall influence of
these elements in the multiple projects.
Table 7 Overall Weight of Each Element for bridges in
Package A to E
The elements having high ratios (or high presence)
were high impact causers. In this multiple project the
overall influence of the elements “Piling” and “Beams”
in bridge completion are critical. Even though the
quantum of these elements were less compared
to other elements of the bridges, the ratio of their
influence in the construction was more due to their
level of special technology, specialist availability,
method of construction, risk involved and limited
usage/availability of resources to produce.
Piling has 25.87% and beams have 38.46% – they have
the maximum presence. The presence of these two
items were more with less volume of work because of
its speciality and use of uncommon materials.
Hence, extra care was given while deciding the
design for these critical elements. Then, few design
optimisations were adopted to reduce the complexity
and to ease implementation in the field as shown
in Table 8. The techniques adopted in element
optimisation for the multiple bridge construction
were successful and resulted in the projects being
executed in time and within budget.
It was observed that the critical elements that needed
more attention were “the foundation (piling) works”
and “Superstructure Beam works”.
82
Package D
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: SYSTEM OPTIMISATION
Table 8. Overall “Piling” and “Beam” element optimisation summary for Project 1 to 5.
FINDINGS AND LESSONS LEARNED
The following findings were obtained and lessons
learnt from the case study:
1. The natural tendency for more cautiousness/
attention for Girders selection was advantages in
the handling and construction of all the beams
(about 355 beams of five different varieties)
without major problems. This enhanced the
finishing of those beams as scheduled in the
bridge programme.
2. Conversely, piling works on the foundation part
were taken lightly at the design stage. There
was no attempt to rationalise and the decision
was left to the individual designers. This resulted
in usage of the same micropile method in the
majority of the bridges. Difficulties which arose
in implementing the micropiles for many bridges
were:(i) Availability of specialists to perform the works
was very limited in Sabah
(ii) The risk of loosing equipment in the drilled
holes requires skilled operators for drilling the
VOLUME Six NUMBER two july - december 2008 PLATFORM
83
Technology Platform: SYSTEM OPTIMISATION
pile. There were not enough skilled operators
in this field.
(iii) The component materials like API pipe G80
and permanent casings were very scarce to
obtain/procure.
(iv) The method had various stages to complete
one point, which is time consuming for each
bridge.
Hence remedial measures were taken to catch up
with the progress at the cost of spending extra
resources, time and money.
CONCLUSIONS
1. In general, the Element Optimisation Technique
was needed to be adopted for all the elements
in compliance with the required standards. Extra
importance is required for elements that have
more influence in the execution and completion
of the project.
2. In multiple DB projects the element optimisations
have to be started well ahead during the
conceptual design stage. But this optimisation
should not interfere in the functional quality and
the integrity of the structure which are designed
for a stipulated life period.
3. In this process it is also important to consider and
review the feedback from field personnel after
conducting a proper study on site conditions.
4. In spite of the lag in piling works as mentioned
in “lessons learned” from the case study, the
company was able to make up and complete
in time with recognition and credits due to the
proper and timely planning in rest of the element
optimisations.
84
ACKNOWLEDGEMENTS
The authors would like to thank the Directors of Kumpulan Liziz
Sdn Bhd., Hj. Ghazali Abd Halim, Mr. Simon Liew and Mr. Liew Ah
Yong for making available data regarding the multiple projects
being executed by the company in Sabah.
REFERENCES
[1]
Rowings, J. E., Harmelink, D. J., & Buttler, L. D., 1991,
“Constructability in the Bridge Design Process”, Research
project 3193, Engineering Research Institute, Iowa State
University. 1-11
[2]
Wikipedia, 2007, Area of Sabah State
Narayanan Sambu Potty received his
Bachelor of Technology in Civil Engineering
from Kerala University and Master of
Technology degree from National Institute
of Technology in Kerala. His PhD work
“Improving
Cyclone
Resistant
Characteristics of Roof Cladding of
Industrial Sheds” was done at Indian
Institute of Technolgy Madras India.
Currently Associate Professor at UTP, he has earlier worked in
Nagarjuna Steels Ltd., TKM College of Engineering, Kerala, India
and Universiti Malaysia Sabah. His main research areas are steel
and concrete structures, offshore structures and construction
management.
C.T. Ramanathan has more than 15 years
of experience in various construction
projects in India and Malaysia. As Project
Manager for Sabah, East Malaysia, he leads
a team of professionals in managing the
Design and Build Infrastructural projects for
this region. He has full responsibility
throughout the D&B project life cycle (initial
project definition to close out) involving
Highways, Roads and Bridges in the state. He also has extensive
experience in the entire process of D&B operation methods. His
Master’s research was on “Complexity Management of Multiple
D&B Projects in Sabah”. He has published and presented ten
papers for international conferences.
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: SYSTEM OPTIMISATION
A Simulation Study on Dynamics
and Control of a Refrigerated Gas Plant
Nooryusmiza Yusoff*, M. Ramasamy and Suzana Yusup
Universiti Teknologi PETRONAS, 31750 Tronoh, Perak Darul Ridzuan, Malaysia
*[email protected]
Abstract
Natural gas has recently emerged as an important source of clean energy. Improving operational efficiency of a
gas processing plant (GPP) may significantly increase its profit margin. This proves to be a challenging problem
due to the time-varying nature of the feedstock compositions and business scenarios. Such fluctuations in
operational and economic conditions are handled by employing advanced process control (APC). Reasonably
accurate steady-state and dynamic simulation models of the actual plant are required for the effective
implementation of APC. This paper deals with the development of such models. Here, a refrigerated gas plant
(RGP), which is the low temperature separation unit (LTSU) of the GPP, is modeled under HYSYS environment.
Calibration and validation of the steady-state model of the RGP are performed based on the actual plant data. A
dynamic model is also built to act as a virtual plant. Main control philosophies of the virtual plant are reported in
this paper. This virtual plant model serves as a platform for performing APC studies.
Keywords: Gas processing plant, dynamic, simulation, control
Introduction
The operators of gas processing plants (GPPs) face
many challenges. At the plant inlet, several gas streams
with different ownerships are mixed as a feedstock to
the GPP. As a result, the feedstock compositions vary
continuously. The feedstock comes in two types: 1)
feed gas (FG) and 2) feed liquid (FL). The FG may be
lean or rich depending on the quantity of natural gas
liquids (NGLs). Lean FG is preferable if sales gas (SG)
production is the main basis of a contract scenario.
On the other hand, rich FG and higher FL intake
may improve the GPP margin due to the increase in
NGLs production. However, this comes at the cost of
operating a more difficult plant. The GPP needs to
obtain a good balance between smooth operation
and high margin.
At the plant outlet, the GPP encounters a number of
contractual obligations with its customers. Low SG
rate and off-specification NGLs will result in penalties.
Unplanned shutdown due to equipment failure may
also contribute to losses. These challenges call for
the installation of advanced process control (APC). A
feasibility study on the APC implementation may be
performed based on a dynamic model. This model
acts as a virtual plant. This way, the duration of the
feasibility study will be shortened and the risk of
interrupting plant operation will be substantially
reduced.
Simulation based on the first principles steady-state
and dynamic models have been recognised as a
valuable tool in engineering. Successful industrial
applications of the first principles simulation are
aplenty. Alsop and Ferrer (2006) employed the
This paper was presented at the 5th International Conference on Foundations of Computer-Aided Process Operations (FOCAPO), Massachusets,
29 June - 2 July 2008
VOLUME Six NUMBER two july - december 2008 PLATFORM
85
Technology Platform: SYSTEM OPTIMISATION
dynamic model of a propane/propylene splitter in
HYSYS to skip the plant step testing completely.
DMCPlus was used to design and implement the
multivariable predictive control (MPC) schemes in
the real plant. In another case, Gonzales and Ferrer
(2006) changed the control scheme of depropanizer
column based on a dynamic model. The new scheme
was validated and found to be more suitable for APC
implementation in the real plant. Several refinery cases
are also available. For examples, Mantelli et al. (2005)
and Pannocchia et al. (2006) designed MPC schemes
for the crude distillation unit (CDU) and vacuum
distillation unit (VDU) based on dynamic models.
The previous works above demonstrate that dynamic
simulation models can be utilized as an alternative to
the empirical modeling based on step testing data.
This way, the traditional practice of plant step testing
to obtain the dynamic response may be circumvented
prior to designing and implementing MPC. Worries
about product giveaways or off-specifications may
then be a thing of the past. In addition, the first
principles simulation model may also be utilised to
train personnel and troubleshoot plant problems
offline. In the current work, the steady-state and
dynamic models of an actual refrigerated gas plant
(RGP) are developed under HYSYS environment. The
steady-state model is initially calibrated with the plant
data to within 6% accuracy. This is used as a basis for
the development of the dynamic model of the RGP,
which is the main objective of this paper. Finally, the
regulatory controllers are installed to stabilize the
plant operation and to maintain product quality.
Simulation
The refrigerated gas plant (RGP) is simulated under
HYSYS 2006 environment (Figure 1). The feed comes
from three streams 311A, 311B and 311C with the
following compositions (Table 1).
The thermodynamic properties of the vapors and
liquids are estimated by the Peng-Robinson equation
of state. Details of the process description have been
described by Yusoff et al. (2007). The steady-state RGP
model has high degree of fidelity with deviation of
less than 6% from the plant data. This is an important
step prior to transitioning to dynamic mode.
Once the steady-state model is set up, three additional
steps are required to prepare the model for dynamic
simulation. The steps are equipment sizing, specifying
Figure 1. HYSYS process flow diagram of the RGP. Equipment’s abbreviation: E=heat transfer unit (non-fired); S=separator;
KT=turboexpander; K=compressor; P=pump; C=column; JT= Joule-Thompson valve; RV=relief valve. Controller’s abbreviation: FC=flow;
TC=temperature; LC=level; SRC=split range; SC=surge; DC=digital on-off; SB=signal block.
86
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: SYSTEM OPTIMISATION
where,
F = mass flow rate (ton∙h−1)
k = conductance (ton∙h−1 ∙ bar−0.5)
∆P = frictional pressure loss as calculated by DarcyWeisbach equation (bar)
Table 1. Compositions of feed gas at different streams
Component
311A
311B
311C
Methane
0.8865
0.7587
0.6797
Ethane
0.0622
0.0836
0.1056
Propane
0.0287
0.0535
0.0905
i-Butane
0.0102
0.0097
0.0302
n- Butane
0.0059
0.0194
0.0402
i-Pentane
0.0003
0.0058
0.0121
n-Pentane
0.0002
0.0068
0.0101
n-Hexane
0.0001
0.0002
0.0028
Nitrogen
0.0039
0.0043
0.0012
Carbon Dioxide
0.0020
0.0580
0.0276
pressure-flow relation at the boundary streams and
installing the regulatory controllers. In the first step,
all unit operations need to be sized accordingly. Plant
data is preferable to produce a more realistic dynamic
model. In HYSYS, an alternative sizing procedure may
be used. Vessels such as condensers, separators and
reboilers should be able to hold 5-15 minutes of liquid.
The vessel volumes may be quickly estimated by
dividing the steady-state values of the entering liquid
flow rates from the holdup time.
For a column, only the internal section needs to be
sized. This is accomplished by specifying the tray/
packing type and dimensions. The tray must be
specified with at least the following dimensions: 1) tray
diameter, 2) tray spacing, 3) weir length, and 4) weir
height. For a packed column, there are a number of
packing types to choose from. Most of the packing
comes with the pre-specified properties such as void
fraction and surface area. The minimum dimension
need to be entered is the packing volume or packing
height and column diameter.
For heat exchangers, each holdup system is sized
with a k-value. This value is a constant representing
the inverse resistance to flow as shown in Equation 1
(Aspentech, 2006):
F = k ∆P
(1)
The k-value is calculated using the converged solution
of the steady-state model. For practical purposes,
only one heat transfer zone is required for the simple
cooler (E-102) and the air-cooler (E-106). The E-102
is supplied with the duty obtained from the steadystate model and E-106 is equipped with two fans. Each
fan is designed to handle 3600 m3/h of air at 60 rpm
maximum speed. The simulation of cold boxes is
more challenging as they are modeled as LNG heat
exchangers. Sizing is required for each zone and layer.
Again for practical purposes, the number of zones in
the cold boxes is limited to three. In each zone, the
geometry, metal properties and a few sets of layers
must be specified. The overall heat transfer capacities
(UAs) and k-values are estimated from the steady-state
solution and specified in each layer.
Equipment with negligible holdup is easily modeled.
For example, the dynamic specification for a mixer is
always set to ‘Equalize all’ to avoid backflow condition.
The flows of fluid across valves are governed by an
equation similar to Equation 1. Here, the k-value is
substituted with the valve flow coefficient, C v. The
valve is sized with a 50% valve opening and 15-30 kPa
pressure drop at a typical flow rate.
The rotating equipment such as pumps, compressors
and expanders may be simulated in a rigorous
manner. The main requirement is the availability of
the characteristic curve of the individual equipment.
The characteristic curves need to be added at the
‘Rating-Curves’ page in order to enhance the fidelity of
the dynamic model. An example of the characteristic
curves for K-102 is illustrated in Figure 2.
The compressors and turbo expanders may also
be modeled based on the steady-state duty and
adiabatic/polytropic efficiency specifications. For
pumps, only the power and static head are required.
This simplification is recommended to ease the
VOLUME Six NUMBER two july - december 2008 PLATFORM
87
Technology Platform: SYSTEM OPTIMISATION
10000
100
7072 RPM
9000
80
4715 RPM
Operating Point
70
Efficiency (%)
7000
Head (m)
90
5725 RPM
8000
6000
5000
4000
60
50
40
3000
30
2000
20
1000
10
0
4000
6000
8000
10000
12000
14000
0
4000
16000
7072 RPM
5725 RPM
4715 RPM
Operating Point
6000
Actual Flow (m3/h)
(a) Head curve
8000
10000
12000
14000
16000
Actual Flow (m3/h)
(b) Efficiency curve
Figure 2. The K-102 characteristic curves
transitioning from steady-state to dynamic mode or
to force a more difficult model to converge easily. In
the current work, K/KT-101 are modeled in this manner
since their characteristic curves are unavailable. In
contrast, K-102, P-101 and P-102 models are based on
the characteristic curves.
The second step in developing a dynamic model is
to enter a pressure or flow condition at all boundary
streams. This pressure-flow specification is important
because the pressure and material flow are calculated
simultaneously. Any inconsistency will result in
the ‘Integrator’ failure to converge. In addition, the
compositions and temperatures of all feed streams
at the flowsheet boundary must be specified. The
physical properties of other streams are calculated
sequentially at each downstream unit operation
based on the specified holdup model parameters (k
and C v). In the current simulation work, all boundary
streams are specified with pressure values. The feeds
enter the RGP at 56.0 barg. The exiting streams are
the NGLs stream at 28.5 barg and the SG stream at
33.5 barg. The flow specification is avoided since the
inlet flows can be governed by PID controllers and
the outlet flows are determined by the separation
processes in the RGP.
The final step is the installation of regulatory
controllers. In most cases, base-layer control is
sufficient. However, more advanced controllers such
as cascade, split range and surge controllers are also
88
installed to stabilize the production. The discussion
of the plant control aspects is the subject of next
section.
Control Philosophies
The control of plant operations is important to meet
product specifications and to prevent mishaps. Due
to space constraints, only main control philosophies
are discussed in this section. The first loop is the TC101 PID controller as shown in Figure 1. The purpose
of this loop is to control Stream 401 temperature
and indirectly the plant overall temperature. This is
accomplished by regulating the cooler duty, E-102Q
through a direct action mode. Stream 401 temperature
is set at 37 °C to prevent excessive FG condensation
into the first separator (S-101).
In the event of excessive condensation, the S-101 level
is regulated by a cascade controller. The primary loop
(LC-101) controls the liquid level in the separator, which
is normally set at 50%. This is achieved by sending a
cascade output (OP) to the secondary loop (FC-102).
The FC-102 acts in reverse mode to regulate Stream
402 flow between 0 and 60 ton/h.
The magnitude of RGP throughput is controlled by
the pressure of second separator (S-102). The S-102
pressure is regulated by the inlet vanes of KT-101 and
the by-pass Joule-Thompson (J-T) valve as per split
range operation (SRC-101). For example, the S-102
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: SYSTEM OPTIMISATION
Feed Gas
pressure is remotely set by FC-101 to decrease from
52.1 to 51.7 bar (SRC-101 PV) in order to increase the
throughput from 280 to 300 ton/h (FC-101 SP). At
the same time, the SRC-101 OP increases from 37 to
40% opening up the KT-101 inlet vanes to 80% while
keeping the J-T valve fully close. The schematic of this
loop is illustrated in Figure 3, the SRC-101 split range
set up in Figure 4 and the close-loop response to a
throughput step up in Figure 5.
E-102
E-103
E-101
S-101
FC
SRC
101
101
Another important controller is PC-101, which
regulates the C-101 top pressure (Ptop) by varying the
K-102 speed. This loop provides the means for
changing the RGP operation from ethane (C2)
to propane (C3) recovery mode or vice-versa.
An example of the close-loop response to the
change from C3 to C2 mode is shown in Figure
6. The C-101 overhead is initially at 24 barg and
-83.0 °C. To recover more C2, the column pressure
is decreased to 22 barg. If all else remain the same,
this action causes the C-101 top temperature (Ttop) to
reduce to -86.4 °C and C2 composition in the SG to
reduce from 2.72 to 2.35 mole %. The downstream
effects can be seen at both suction and discharge
sides of K-102. Lowering the C-101 Ptop will reduce the
K-102 suction pressure (Psuct) from 25.0 to 23.5 bar.
Consequently, the K-102 speed needs to be increased
from 5 684 to 6 352 RPM in order to obtain a discharge
pressure of 33.7 bar. The increase in K-102 speed causes
its discharge temperature (Tdisc) to also increase from
40.6 to 60.2 °C.
KT-101
To C-101
J-T
S-102
Figure 3. Plant throughput control
100
OP (%)
80
60
KT-101
J-T valve
40
20
0
0
4
8
12
16
20
SRC-101 Signal Output (mA)
300
52
51. 9
290
51. 8
SRC-101 PV
285
280
275
0.0
51. 7
FC-101 SP
51. 6
10. 0
20. 0
30. 0
Time (min)
40
39. 5
SRC-101 O P
295
40. 5
40. 0
50. 0
51. 5
60. 0
39
38. 5
38
37. 5
(%)
52.1
SRC-101 OP
305
SRC-101 PV (bar)
FC-101 SP (ton/h)
Figure 4. SRC-101 split range setup
37
36. 5
Figure 5. Response to throughput step up
Figure 6. Effect of C-101 Top Pressure
VOLUME Six NUMBER two july - december 2008 PLATFORM
89
Technology Platform: SYSTEM OPTIMISATION
Figure 7. Effect of TC-101 failure on C1 composition in SG
The importance of regulatory control in meeting
product specification is illustrated in Figure 7. Here,
the methane composition in feed changes from
88.7 to 80.6% at the time when TC-101 is offline but
other controllers remain online. As a result, stream
401 temperature (TC-101 PV) increases from -38.0 to
-23.8 °C. This causes the C-101 top temperature to
increase from -86.3 to -66.9 °C. A hotter top section
of C-101 induces methane losses to the bottom of the
column as confirmed by the decrease in SG methane
composition from 96.9 to 91.5%, or by 5.4%.
Conclusions
A simulation model of refrigerated gas plant (RGP) was
successfully developed. The dynamics of the RGP was
simulated based on a high fidelity steady-state model.
Controllers are installed to regulate plant operation
and to stabilize production. The dynamic model can
be used as a virtual plant for performing advanced
process control (APC) studies.
References
[1]
Alsop N. and Ferrer J. M. (2006). “Step-test free APC
implementation using dynamic simulation”. AIChE Spring
National Meeting, Orlando, Florida, USA
[2]
Aspentech (2006). “HYSYS 2006 Dynamic Modeling Guide”.
Aspen Technology: Cambridge, MA, USA
[3]
Gonzales R. and Ferrer J. M. (2006). “Analyzing the value
of first-principles dynamic simulation”. Hydrocarbon
Processing. September, 69-75
90
[4]
Mantelli V., Racheli M., Bordieri R., Aloi N. Trivella F. and
Masiello A. (2005). “Integration of dynamic simulation and
APC: a CDU/VDU case study”. European Refining Technology
Conference. Budapest, Hungary.
[5]
Pannocchia G., Gallinelli L., Brambilla A., Marchetti G. and
Trivella F. (2006). “Rigorous simulation and model predictive
control of a crude distillation unit”. ADCHEM. Gramado,
Brazil.
[6]
Yusoff N., Ramasamy M. and Yusup S. (2007). “Profit
optimization of a refrigerated gas plant.” ENCON. Kuching,
Sarawak, Malaysia.
Nooryusmiza Yusoff graduated from
Northwestern University, USA with BSc in
Chemical Engineering and subsequently
became a member of the American
Chemical Engineering Honors Society
“Omega-Chi-Epsilon”. He received the MSc
Degree from the University of Calgary,
Canada with a thesis on “Applying
Geostatistical Analyses In Predicting Ozone
Temporal Trends”. He is currently pursuing his PhD at Universiti
Teknologi PETRONAS (UTP) after serving as a lecturer in the
Department of Chemical Engineering, UTP. His areas of research
interest mingle around process modeling and simulation,
advanced process control as well as multiscale system.
M. Ramasamy is presently an Associate
Professor in the Department of Chemical
Engineering at Universiti Teknologi
PETRONAS (UTP). He graduated from
Madras University in the year 1984. He
obtained his masters degree with
specialisation in Process Simulation,
Optimisation and Control from the Indian
Institute of Technology, Kharagpur in 1986
followed by his PhD in 1996. His areas of research interests include
modeling, optimisation and advanced process control. Dr
Ramasamy has guided several undergraduate and postgraduate
students and published/presented several technical papers in
international refereed journals, international and national
conferences. He has delivered a number of special lectures and
also successfully organised seminars/symposiums at the national
level. Presently, he is the project leader for the PRF funded project
on “Optimisation of Energy Recovery in Crude Preheat Train” and
the e-Science funded project on “Soft Sensor Development”.
Suzana Yusup is presently an Associate
Professor at Universiti Teknologi PETRONAS
(UTP).
She graduated in 1992 from
University of Leeds, UK with an Honours
BEng. Degree in Chemical Engineering. She
completed her MSc in Chemical Engineering,
(Adsorption) in 1995 and PhD in Chemical
Engineering in Powder Technology at
University of Bradford UK in 1998. Her main
research areas are in the area of adsorption and reaction
engineering. She was lead researcher for natural gas storage of
direct fuel NGV Engine Injection under IRPA project and a research
leader for a cost effective route for methanol production from
natural resources via liquid phase synthesis and synthesis of
carbon nano-fibres and tubes for hydrogen storage under
e–science project.
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
AN INTERACTIVE APPROACH TO CURVE FRAMING
Abas Md Said*
Universiti Teknologi PETRONAS, 31750 Tronoh, Perak Darul Ridzuan, Malaysia
*[email protected]
Abstract
Curve framing has numerous applications in computer graphics. Commonly used curve framing techniques such
as the Frenet frame, parallel transport frame and ‘up-vector’ frame cannot handle all types of curves. Mismatch
between the technique used and the curve being modeled may result in the extruded surfaces being twisted
or an object erratically rotating while flying. We propose an interactive approach to curve framing. The user
selects a ‘twisted’ portion of the curve and untwists it accordingly, as required by its specific application needs.
The technique works fine as per user’s wish.
Keywords: Curve framing, Frenet frame, parallel transport frame.
INTRODUCTION
Curve framing is the process of associating coordinate
frames to each point on a three-dimensional space
curve. This can be depicted as in Figure 1, where a
number of local Cartesian coordinate frames are
drawn on a curve. This is very useful when one needs
to treat a coordinate locally, rather than globally. It has
numerous applications in computer graphics, such as
in the construction of extruded surfaces (Figure 2),
providing the orientation of flying objects (Figure 3)
and visualisation during a fly-through.
Figure 2: Extruded surface from a curve
Figure 1: Some local frames on the spine
Figure 3: Flying along a 3D path
This paper was presented at the International Conference on Mathematics & Natural Sciences, Bandung,
28 - 30 October 2008
VOLUME Six NUMBER two july - december 2008 PLATFORM
91
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
In these applications, frame orientations are essentially calculated based on the curve itself. A few
approaches commonly employed in these applications
are the Frenet, parallel transport, ‘up-vector’ and
rotation minimising frames [6, 7]. Having figured
out the frames, the necessary points (or vectors) to
construct the surface or orient an object can be easily
obtained. While these approaches have been very
useful, not all can successfully address all types of
curves. At times, the inherent nature of a curve causes
‘twists’ to occur, giving distorted extruded surfaces or
an object flying in a twist. This paper looks at some
of the problems with some of these approaches and
proposes an interactive technique to address the
issue.
Figure 4 (a)
THE FRENET FRAME
A Frenet frame for a curve C(t) is formed by a set of
orthonormal vectors T(t), B(t) and N(t) where
T(t ) =
C ' (t )
C ' (t )
is tangential to C(t),
B(t ) =
C ' (t ) × C ' ' (t )
C ' (t ) × C ' ' (t )
is the binormal and
N(t ) = B(t ) × T(t ) Figure 4 (b)
is the “inner” normal,
i.e., on the osculating circle side of the curve [2, 5]
(Figure 4a). A portion of the extruded surface of
a curve is depicted in Figure 4. Depending on the
curve, it can be seen that at some point on the curve,
i.e., near the point of inflection, the frame starts to
twist (Figure 4b, C (t ) = (cos 5t , sin 3t , t ) ), while on
another, there are no twists at all (Figure 4c,
C (t ) = (cos t , sin t , 0) ). The twist occurs because the
inner normal now starts to “change side”.
Another problem may also occur when the second
derivative of C (t ) vanishes, i.e., on a momentary
straight line [4]. This is clear from vector B(t ) as
C ' ' (t ) would be identically 0 for a straight line.
92
Figure 4 (c)
Figure 4: A Frenet frame (a) and extruded surfaces by Frenet
frames, twisted (b) and untwisted (c)
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
Figure 5: Parallel curves
PARALLEL TRANSPORT FRAME
Parallel transport refers to a way of transporting
geometrical data along smooth curves. Initial vectors
u and v, not necessarily orthogonal, are transported
along the original curve by some procedure to generate
the other two parallel curves as in Figure 5 [4]. Hanson
and Ma [4] showed that the method reduced twists
significantly. In another similar application, Bergou
et al. successfully used parallel transport frames in
modeling elastic discrete rods [1]. Despite generating
good extruded surfaces, the method may not meet
requirements for a fly-through.
Figure 6a
Figure 6b
‘UP-VECTOR’ FRAME
Another method that could be applied to avoid twists
is to use the ‘up-vector’ [3]. In this approach, view the
frames similar to Frenet’s, with Ti ’s, B i ’s and N i ’s,
moving along the curve. Without loss of generality,
assume y-axis as the global vertical coordinate. Then
find the unit ‘radial’ vector in the unit circle (spanned
by the vectors B i ’s and N i ’s) that gives the highest
y-coordinate value (Figure 6a). The frame is then
rotated about Ti until N i coincides with the ‘radial’
vector.
Figure 6c
While this approach works on many curves, e.g., as
in Figure 6b, it fails when the curve is momentarily
vertical or abruptly tilts the opposite ‘side’ (Figures 6c
and d).
Applications-wise, this technique corrects the twists in
extruded surfaces or a walkthrough for many curves.
However, it may not be realistic when simulating fastmoving objects, e.g., airplanes, due to issues such as
inertia and momentum during flight [8].
Figure 6d
Figure 6: ‘Up-vector’ construction and its problems
VOLUME Six NUMBER two july - december 2008 PLATFORM
93
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
Figure 7: Roll
Figure 8: Generating the frames
UNTWISTING THE TWISTS, INTERACTIVELY
In this study’s approach, a simple method employed
was by iteratively generating the frames based on
any three successive points on the curve (Figure 8)
and applying it on a plane flight trajectory. Given
three points A, B and C, the normal vector to the
plane parallel to ∆ABC, q, is easily calculated using
the coordinates of the points. The tangential vector
is crudely approximated by p = C – A and the ‘inner’
normal r by the cross product q × p. Following this
method, the frames were still subject to twists.
Twists were eliminated by employing the ‘up-vector’
procedure (Figure 9). The choice for this was fairly
clear, as in most applications an object set to fly will
begin in an up-right position. Furthermore it was a
better perspective to set the twist. Looking at the
whole picture of the curve, the where and how to
twist or untwist the curve (Figure 9) were identified.
The corresponding plane orientation with respect to
the extruded surface in Figure 9 is shown in Figure
10. Upon identification of the intervals, each could be
twisted and untwisted piece-wise. At position I, for
example, the plane should have more roll compared
to that in Figure 10 due to the plane speed and path
curvature.
Figure 9: ‘Up-vector’ results
Figure 10: Flight orientation (‘up-vector’)
Despite good extruded surfaces produced by parallel
transport and ‘up-vector’ frames, the techniques may
not have intended effects for graphics applications
such as a fly-through. In such a case, user interactions
are desirable to make rolling more reasonable during
flights (Figure 7). This is to adjust or correct the object
orientation using the user’s common sense. As such,
one would like to be able to interactively control the
twists.
94
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
REFERENCES
Figure 11: User intervention
The twisting was achieved by rotating the frame for
each increment about the local tangential axis. The
result of such operation is shown in Figure 11, with the
plane flying along the path, rolling according to the
new orientation. It should be noted that the purpose
of the dark band on the extruded surfaces in Figure 9 is
to visualise the twist on the curve while the flight path
is the curve itself (Figures 10 and 11). At the instance
when the dark band is up, the plane is upright, and
when the band is below the curve, the plane is upside
down. Figure 11 shows the plane flying orientation
at positions I, J and K after the user has adjusted the
twists in the curve as per his perception.
[1]
M. Bergou, M. Wardetzky, S. Robinson, S., B. Audoly and E.
Grinspun (2008). Discrete Elastic Rods, ACM Transactions
on Graphics, Vol. 27, No. 3, Article 63.
[2]
R. L. Bishop (1975). There is More than One Way to Frame a
Curve, Amer. Math. Monthly, Vol. 82, No. 3, pp. 246 – 251.
[3]
J. Fujiki and K. Tomimatsu (2005). Design Algorithm of “the
Strings - String Dance with Sound”, International Journal of
Asia Digital Art and Design, Vol. 2, pp. 11 – 16.
[4]
A. J. Hanson and H. Ma, H (1995). Parallel Transport
Approach to Curve Framing, Technical Report 425, Indiana
University Computer Science Department.
[5]
F. S. Hill (2001). Computer Graphics Using OpenGL, Prentice
Hall, New Jersey.
[6]
T. Poston, S. Fang and W. Lawton (1995). Computing and
approximating sweeping surfaces based on rotation
minimizing frames, Proceedings of the 4th International
Conference on CAD/CG. Wuhan, China.
[7]
W. Wang, B. Jüttler, D. Zheng and Y. Liu (2008). Computation
of Rotation Minimizing Frames, ACM Transactions on
Graphics, Vol. 27, Issue 1, pp. 1 – 18.
[8]
J. Wu and Z. Popovic (2003). Realistic Modeling of Bird
Flight Animations, International Conference on Computer
Graphics and Interactive Techniques, ACM SIGGRAPH 2003
Papers, pp. 888 – 895.
Abas Md Said holds BSc and MSc degrees
from Western Michigan University and PhD
from Loughborough University. He is a
senior lecturer at UTP. His research interests
are in computer graphics, visualisation and
networks.
CONCLUSIONS
Most available methods to curve framing yield
acceptable results in generating extruded surfaces.
While parallel transport and ‘up-vector’ frames
generally work well in these cases, they may not be
adequate to model a fast fly-through. This study
proposed an interactive approach to address the
issue. The approach does away with the physics of
flight and relies on user’s insight of flight trajectory.
The approach lets the user fine-tune the way a fastmoving object should roll about its axis. This is
necessary in this kind of applications because unless
the necessary physics is embedded in the model, the
Frenet, parallel transport and ‘up-vector’ methods
previously discussed may overlook the roll problem.
VOLUME Six NUMBER two july - december 2008 PLATFORM
95
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
Student Industrial Internship Web Portal
Aliza Sarlan*, Wan Fatimah Wan Ahmad, Dismas Bismo
Universiti Teknologi PETRONAS, 31750 Tronoh, Perak Darul Ridzuan, Malaysia
*[email protected]
Abstract
A Student Industrial Internship Web Portal (SIIWP) was developed to automate the current manual business
processes. The portal allowed internship eligibility checking, registration, student-lecturer assignment, visit
schedule, online-logbook submission and monitoring as well as grade book of industrial internship programme
at Universiti Teknologi PETRONAS. PHP 5, Easy PHP 2.0, Macromedia Dreamweaver MX 2004, MySQL Database
and Apache Web Server were used to develop SIIWP. Phased development model was used in the development
process by applying business process improvement technique. Findings showed that the prototype could be
used as a communication medium for all parties involved during the industrial internship programme. The
System could be easily used as an aid for the internship programme.
Keywords: Industrial internship, web portal, business process improvement, phased development model, business
process
Introduction
Student Industrial Internship Programme (SIIP) is part
of the curriculum for most higher learning institutions
worldwide. Its main purpose is to expose students
to a real working environment and relate theoretical
knowledge with applications in the industries. The
objectives are to produce well-rounded graduates
who possess technical competence, lifetime learning
capacity, critical thinking, communication and
behavioural skills, business acumen, practical aptitude
and solution synthesis ability [1]. Issues such as, long
distance learning, communication, monitoring and
management arise as crucial to ensure the success of
the programme.
The SIIP is a thirty-two-week programme where
Universiti Teknologi PETRONAS (UTP) students
are attached to various companies in or outside
Malaysia. The purpose of SIIP is to expose UTP
students to the world of work, thus they are able
to relate theoretical knowledge with application
in the industry. Furthermore, SIIP can enhance the
relationship between UTP and the industry and/or the
government sector. The Student Industrial Internship
Unit (SIIU) is responsible for handling and monitoring
the process of students’ internship in UTP. SIIU is
responsible for internship processes from student
application to checking students’ eligibility status,
internship placement application and confirmation,
lecturer visit scheduling and grading. Many problems
arise since all processes are still been done manually,
such as missing data and redundancy, delay in the
grading process, communication problems and most
crucial is student monitoring. Currently, by telephone
and email are the main methods of communication
which have imposed many problems, e.g. update to
all students has to be approached individually and
resulted in high cost of communications.
This paper was presented at the International Symposium on Information Technology 2008, Kuala Lumpur
26 - 29 August 2008
96
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
The main objective of this project is to develop a
prototype of Student Industrial Internship Web Portal
(SIIWP) that automates current manual processes to
reduce possible problems in communication, data
loss and redundancy. It makes monitoring, instructor
assignment and scheduling, grading and reporting
easy and to a greater extent, error free.
Related Works
Distance Learning
The SIIP refers to the placement of students on a
spread of locations for implementing their knowledge
or having practicals in the real industry sector [2].
Thus, the distance learning concept is actually being
implied in the context of SIIU.
Distance Learning and the WWW Technologies
The popularity of the World Wide Web (WWW)
has made it a prime vehicle for disseminating
information[3]. The advance of WWW technologies
has driven the usage of Internet to new applications
and at an unprecedented rate. Education institutions
have/had developed and/or migrated to web-based
applications in overcoming the distant learning issue
and to provide better services for their users, i.e.
students and teachers.
Web-based System
A web-based application is an application that is
accessed with a web browser over a network such
as the Internet or an intranet [4]. The web-based or
automated system provides far more efficiency in
processing any task domain especially for a system
that involves a lot of data collection and retrieval [5].
Web-based systems should meet its stakeholders’
(users’) requirements and expectations. Thus, webbased applications should be developed on-top of a
carefully studied business process of the organisation
in which it is to be deployed.
The research project will apply a similar strategy, that
is, to create a web-based system built based on the
current business process of SIIU. A web portal would
be the most appropriate web-based system for SIIU.
Web Portals
A web portal serves as a starting point when users
connect to the Internet, just like a doorway to web
services that guide users to the right information they
need[6]. A web portal is defined as a site that provides
a starting point for users to explore and access
information on the WWW or intranet. This starting
point could be a general purpose portal or it could be
a very specialised portal, such as a homepage [7].
According to Strauss [8], a common way is to divide
web portals into two categories:
a) Horizontal Portal – A horizontal portal is an
internet portal system that is open to the public
and is often considered as a commercial site.
Most horizontal portals offer on a single web page
a broad array of resources and services that any
user may need. Examples of horizontal portals:
Yahoo!, MSN.
b) Vertical Portal – A vertical portal is a web site that
provides information and services as defined and
requested by users.
Strauss also mentioned that a true portal should be:
1) Customised – A true portal is a web page whose
format and information content are based on
information about the user stored in the portal’s
database. When the user authenticates (logs in)
to the portal, this information determines what
the user will see.
2) Personalised – The user can select and store a
personal set of appearance content characteristics
for a true portal. These characteristics may be
different for every user.
3) Adaptive – The portal gets to “know” the user
through information the user supplies and through
information the portal is programmed to gather
about the user. As the user’s role in the institution
changes, a true portal will detect that change and
adapt to it without human intervention.
VOLUME Six NUMBER two july - december 2008 PLATFORM
97
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
4) Desktop-Oriented – The goal of a portal is to mask
the inner workings of the campus information
systems from the user. Signing-on into the portal
keeps a user from having to sign each of the many
systems, on campus and off, that provide the
portal content. The ultimate portal could become
the user’s point of entry not just into campus and
internet web spaces, but also into his or her own
desktop computer.
The web portal developed for SIIU falls into the second
category – vertical portal. The web portal allowed
users to access related information with customised
functionalities that meet the users’ requirements.
Web portals have been a successful support system
in education institutions. Thus, many education
institutions have/had been implementing vertical
portals:
Business Process
A Business Process is a complete and dynamically
coordinated set of collaborative and transactional
activities that deliver value to customer [10]. It is
focused upon the production of particular products;
these may be physical products or less tangible one,
like a service [11]. A Business Process can also be
defined in a simpler and more general description
as a specific ordering of work activities across time
and place, with a beginning, and an end, and with
clearly identified inputs and outputs – a structure for
action [12]. Essentially, there are four key features to
any process (i) predictable and definable inputs, (ii)
a linear, logical sequence or flow, (iii) a set of clearly
definable tasks or activities, and (iv) a predictable and
desired outcome or result.
Business Process Improvement (BPI)
1) The Universiti Teknologi PETRONAS Information
Resource Centre implemented a web-based
system with an objective to ease the users’ task
in finding the resources based on an indexed
location and the staffs’ tasks by having an
organised way of indexing. The main reason why
it is made available online is that its database is
located in a remote location from its users. The
system’s functionalities are to search for the
available resources and their indexes, to reserve
discussion rooms, and to search for online journals
from online journal providers [5].
2) Indiana University is a large and complex
institution consisting 8 campuses, over 92,000
students, and 445,000 alumni. The university
created an enterprise portal to which faculty,
staff, students, alumni, prospective students,
and others will travel and uncover a broad array
of dynamic web services. “OneStart” (http://
onestart.iu.edu) provides a compelling place for
faculty, staff, and students to find services with
the intent of developing life-long campus citizens
of the university [9].
98
The Business Process Improvement (BPI) method
is used to identify the requirements of the ‘tobe’ system. BPI has been used in developing and
implementing the prototype system. BPI is defined
as making moderate changes to the way in which
the organization operates to take advantage of new
opportunities offered by technology. The objective
of any BPI methodology is to identify and implement
improvement to the process; thus BPI can improve
efficiency and effectiveness[13]. The scope of BPI
compared to Business Process Reengineering (BPR) is
narrow and short term [12].
Industrial Internship Business Process
Industrial Internship business process is divided into
3 major phases namely Pre-industrial Internship;
During-industrial Internship and Post-industrial
Internship. Figures 1, 2 and 3 show the 3 major phases
of the industrial internship business process.
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
Start
Start
SIIU give briefing to students
Student submit logbook
Students submit form to check eligibility
Lecturers collect students’ Final
Report & logbook From SIIU
Not
eligible
SIIU check the
II coordinators key in students’
marks and submit to SIIU
Student register for internship
Submit applications and resume to SIIU
& potential host company
Receive offer of placement
Not ok
SIIU check the
marks
The complete marks (grades)
forwarded to exam Unit
Placement offer Notification
End
End
End
Figure 1. Phase 1 pre-internship
Figure 3. Phase 3 post-internship
Start
Problem Identification
Students register at host companies
Students submit form and training
schedule to confirm placement
List of students forwarded to coordinator
Not
ok
SIIU Make
confirmation with
SIIU confirms the visit schedule
SIIU inform respective lecturers of visit
schedule
First & Second visit commence
st
Visiting lecturer submit report for 1 visit
and students’ marks for 2nd visit
End
Currently, the system is working manually. However,
number of problems and pitfalls has been increasingly
arising and causing some deficiencies in the system.
Some of the problems in the current manual system
are:
 Manual and time consuming process of student’s
eligibility status identification due to manual cross
checking process.
 Manual students registration for industrial
internship by filling up a paper form require SIIU
staff to key-in the students particular and contact
details in to the excel spreadsheet. This poses
problems such as data error due to human error
as well as too time-consuming.
 Difficulties
and
ineffective
method
in
communicating with the students since all
communications are by telephone and email.
 Loss of students’ placements applications, resume
and other important documents due to too many
papers and manual process involved.
Figure 2. Phase 2 during-internship
VOLUME Six NUMBER two july - december 2008 PLATFORM
99
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
 Manual system using Microsoft Excel and Word
provide limited features just for entering, searching
and printing the data.
 There is no efficient way to notify the status of
placement, post announcement and update the
placement of internship.
 The manual system only supports two computers
with a centralise database. As a result, it has
limited the ability to access of information.
 All the business processes depends mostly on the
one person who knows the processes. Others
have difficulty to interrupt in order to complete
the process.
 Difficulty in monitoring the students’ progress
and performances as the assigned lecturers to
the lecturers usually receive and view the weekly
report at the end of the programme. Weekly
report submission to UTP lecturers been done
manually by fax or postal mail. Hence, it may
end up missing and may not reach the respective
lecturers on time.
 Grade compilation and calculation that been done
manually by individual lecturer supervisor always
poses problem such as missing evaluation forms
and delays in final grade submission.
Therefore, the new system (SIIWP) needed to be
developed to automate and improve most of the
manual processes to reduce errors and time, and to
increase the efficiency in student monitoring and
communications.
SIIWP Development
The SIIWP automates and improves SIIU’s business
process in conducting SIIP; emphasising on efficiency
and effectiveness. The SIIWP serves the following
objectives:
 To closely monitor the students’ performances by
allowing the SIIU and respective lecturers to view
the students’ weekly reports online and ensure
that the students are monitored closely in a timely
manner.
 To ease the task of scheduling the visits and
assigning UTP Supervisors, and to assist in
100
assigning the students to the respective lecturers,
based on their programme of study and location
of the host companies. Besides, the lecturers
can also view the list of students assigned under
their supervision together with the name of host
companies. SIIU need neither to make calls to
inform the lecturer nor the students about it. This
will reduce the workload of SIIU and make the
process more organised.
 To automatically calculate the final mark of
students at the end of the internship programme
once their marks have been entered into the
system. Besides preventing miscalculations, this
type of task automation actually assists in freeing
staff’s time and workload.
 To generate specific reports based on certain
criterion, such as the Host Company, location,
programme of study and the supervisor for further
references and analysis. The SIIWP is expected to create process improvements
that lead to better effectiveness; thus, the Business
Process Improvement (BPI) analysis technique is used
as the method for requirements analysis in identifying
the critical business processes which need to be
automated. Basically, BPI means making moderate
changes to the way in which the organisation
operates to improve efficiency and improve
effectiveness. BPI projects spend significant time on
understanding the current as-is system before moving
on to improvements and to be system requirements.
Duration analysis activities are performed as one of
the BPI techniques[13]. The improved business
process is more efficient whereby certain processes
can be done concurrently and thus reduce the
duration of the initial manual business process.
The SIIWP complies with the open source standards as
SIIWP is developed using PHP5, XHTML, and JavaScript
for the server-side scripting, Easy PHP 2.0, Macromedia
Dreamweaver MX 2004 as the development tool and
MySQL for the database. The SIIWP is developed
based on the phased development model. Using the
model, the project is divided into small parts. The most
important and urgent features is bundled into the first
version of the system. Additional requirements will be
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
SIIWP Version 1.0:
Pre-Industrial Internship
1. Identify Eligibility Status
2. Register to System
3. Track Application Process
4. Update & Retrieve Database
5. Report Generation
PORTAL
SIIWP Version 2.0:
During-Industrial Internship
1. Confirmation of Placement
2. Schedule Visit
3. Monitor Student – weekly report online
4. Post Announcements
Excel
Reader
Application
Tracker
Student
Monitoring
Grading
Post
Announcement
Company
Resource Centre
DBMS
SIIWP Version 3.0:
Post-Industrial Internship
1. Grade Student
2. View Grade
SIIWP Database
Figure 4. SIIWP Functionalities
Figure 5. SIIWP system architecture
added to the system in the next consecutive versions
of the system. The model allowed the development
team to demonstrate results earlier on in the process
and obtain valuable feedback from system users.
There are 3 system version releases of SIIWP. SIIWP
version 1.0, version 2.0 and version 3.0 concentrates on
pre-internship, during-internship and post-internship
business processes respectively. Each version
encompassed previous requirements and additional
requirements. Figure 4 shows the functionalities of
SIIWP version 1.0, version 2.0 and version 3.0.
staff (administrator), coordinator and lecturer. When
users log on to the system, they will be directed to
their personalised portal.
SIIWP System Architecture
The SIIWP adopted the combination of data-centred
and client-server architectural model. A client-server
system model is organised as a set of services and
associated server(s) and clients that access and use
the services. The server itself, or one of the server(s),
contains a database where data are stored, thus it
adopts data centred architecture. Figure 5 illustrates
the SIIWP system architecture.
Main Index Portal
The Main Index Portal is a point of access for all users
of SIIWP. It provides generic information that all users
can access such as Announcement and Upcoming
Events, Search Companies, About Student Industrial
Internship and Help. All Users can login to access
their personalised portal from the Main Index Portal.
Students are to register themselves to the system
before a personaliszed portal is assigned to them.
Before registering, students are required to check
their eligibility status to undergo the SIIP. Figure 6
illustrates the SIIWP Main Index Portal.
SIIWP Prototype
The SIIWP consists of a main index portal (Main
Portal) and 4 types of user personalised portals,
namely: (i) Admin (Staff) Portal, (ii) Student Portal,
(iii) Lecturer Portal, and (iv) Coordinator Portal. There
are different types of users that can log on to the
system with different level of access – student, SIIU
Figure 6. Main index portal
VOLUME Six NUMBER two july - december 2008 PLATFORM
101
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
Figure 7. Admin staff portal
Figure 8. Student portal
Admin (Staff) Portal
is a personalised point of access for UTP Lecturers. It
contains generic information and functionalities that
are only accessible by Lecturers: Grade Students and
Supervise Students. The Grade Students functionality
enabled lecturers to grade students online by keying
in the student’s grades into the system. The system
can automatically calculate students’ grade. The
grade calculated can be viewed by staff and the
respective student. The Supervise Students function
enabled Lecturers to view all students under his or her
supervision.
The Admin (Staff) Portal is a personalised point of
access for SIIU staff. It contains generic information and
personalised functionalities that are only accessible
by staff: Database Statistical Analysis, Database
Tables and Upload Document. Figure 7 illustrates the
Admin (Staff) Portal.
Student Portal
The Student Portal is a personalised point of access
for students; it contains generic information and
functionalities that are only accessible by students:
Application Status and Weekly Report Submission.
The application status fuctionality allowed students to
register their internship application(s) and update their
application status into the system either: (i) Send CV/
Resume, (ii) Interview, (iii) Offer Letter, (iv) Accepted,
and (v) Rejected. This information was accessible for
viewing by Lecturers and Staff; hence they were able
to monitor and provide assistance. The Weekly Report
Submission functionality enabled students to submit
their report online and to generate (print) it. The
weekly reports that were submitted online could be
viewed by lecturers and staff. This enabled Lecturers
and staff to monitor and know student’s activities in
the host company. Figure 8 illustrates the Student
Portal.
Coordinator Portal
The Coordinator Supervisor Portal is a personalised
point of access for lecturers selected as SIIP
coordinator in their respective department, it contains
generic information and functionalities that are only
accessible by Coordinators i.e. Schedule Lecturer
Visits enabled Coordinators and SIIU staff to detemine
Lecturer Portal
The Lecturer Supervisor Portal as shown in Figure 9
102
Figure 9. Lecturer portal
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
Figure 10. Coordinator portal
Figure 11. Acceptance test result
the appropriate date for Lecturer’s 1st and 2nd visits.
Figure 10 illustrates the Coordinator Portal.
The acceptance test focused on 3 criteria: (i) user
interface, (ii) user friendly navigation, and (iii) basic
functionalities. For the system to pass the acceptance
test, at least 75% of test participants (22 participants)
agreed that the system’s criteria under consideration
met their expectations. The acceptance test concluded
that 90% of the test participants agreed that the user
interface met their expectations, 85% of the test
participants agreed that it had user-friendly navigation,
and 78.3% of the test participants found that the basic
functionalities met with their expectations. Figure 11
charts the acceptance test results.
Testing
To eliminate system faults, fault removal strategy was
implemented in every system version development.
Fault removal is basically a Validation and Verification
(V&V) process; checking and analysis process consists
of requirements reviews, design reviews, code
inspections, and product testing. Table 1 summarises
the testing conducted.
Acceptance testing was conducted for SIIWP System
Version 3.0; 30 users (students, lecturers, and staffs)
participated in the testing. The purpose of the
acceptance testing was to check if the system has met
with the functional and non-functional requirements
defined in the requirements determination phase
and if the system matched exactly or came closely
matching users expectation.
Table 1. System testing
Feature Test
Execution of an operation/
functionality with interaction
between operations minimized.
Load Test
Testing with field data and
accounting for interactions.
Regression Test
Feature test after every build
involving significant change.
Acceptance Test
(Alpha Testing)
Test conducted by users to ensure
they accept the system.
Thus, System Version 3.0 was considered to have
passed the acceptance test. However, users gave
feedback and suggestions of other functionalities that
may be useful to the SIIWP.
Conclusion and Future Work
The SIIWP met all functional and non-functional
requirements. As SIIWP was built specifically on top of
a well studied and improved SIIU business process, its
basic functionalities of the business process matched
closely to users’ expectations. SIIWP was able to
solve the distant learning problem in SIIP. This is
essential to SIIU as it will improve business processes
for SIIU, in conducting SIIP, through automation. The
implementation of SIIWP can help SIIU ensure the
success of SIIP by providing optimal and high quality
service. At the same time, the new system may be
VOLUME Six NUMBER two july - december 2008 PLATFORM
103
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
able to become a central internship and job resource
centre for UTP students.
[5]
Norfadilah Bt. Samsudin (2006). Universiti Teknologi
PETRONAS. Online Industrial Training System, Final year
project report, 1 - 54.
There are still limitations in SIIWP: limited direct
communication between users, potential data
redundancy, and accessibility from outside UTP.
Users also gave feedback and suggestions of other
functionalities that may be useful to the SIIWP.
These suggestions are functionalities for further
enhancements that can be further studied and be
added into the SIIWP. Future work could include:
[6]
Zirpins, C., Weinreich, H., Bartelt ,A and Lamersdor, W. (2001).
Advanced Concepts for Next Generation Portals, 12th
International Workshop on Database and Expert Systems
Applications, 501-506.
[7]
Aragones, A. & Hart-Davidson, W. (2002). Professional
Communication Conference, 2002. IPCC 2002. Proceedings.
IEEE International. Why, when and how do users customize
Web portal?, 375 – 388.
[8]
Strauss, H. (2002). In Web Portals & Higher Education:
technologies to Make IT Personal. All About Web Portals: A
Home Page Doth Not Portal Make, 33-40.
[9]
Thomas, J. (2003). Indiana University’s Enterprise Portal
as a Service Delivery Framework. Designing Portals:
Opportunities and Challanges, 102-126.
Direct Communication Media: Forum, Blogs, Chat
Systems. SIIWP only provides contact details of the
users. Forum, blogs, and chat systems can be added
into the SIIWP a direct communcation media between
users of SIIWP.
Frequently Asked Questions: Frequently Asked
Questions (FAQ) could be added into SIIWP so
that users can share particular knowledge and/or
experience gained through the SIIP, stores them in the
database, and be mapped to certain issues. The FAQ
function enables users to interact and ask questions
to the system regarding specific issues faced during
their SIIP.
Data Mining: An effective data mining technique can
be implemented to SIIWP to provide users with a more
desirable view of the database. References
[1]
Universiti Teknologi
internship guidelines.
PETRONAS, (2000).
Industrial
[2]
Aliza Bt Sarlan, Wan Fatimah Bt Wan Ahmad, Judy Nadia
Jolonius, Norfadilah bt Samsudin (2007). Online Web-based
Industrial Internship System. 1st International Malaysian
Educational Technology Convention 2007. UTM , Johor. 25 Disember 2007, 194-200.
[3]
Liu, Haifeng, Ng, Wee-Keong & Lim, Ee-Peng, 2005.
Scheduling Queries to Improve the Freshness of a Website.
World Wide Web: Internet and Web Information Systems, 8,
61-90.
[4]
Wikipedia®, the free encyclopedia (2007). System. Retrieved
February 12, 2007, from http://en.wikipedia.org/wiki/
System.
104
[10] Smith, H.; Finger, P. (2003) IT doesn’t matter – Business
Process Do. August 2003. Meghan-Kiffer Press 2003.
[11] Aalst, W. & Hee, K. (2002), Workflow Management. Models,
Methods, and Systems, MIT Press.
[12] Davenport, T. 1998. “Some principles of knowledge
management”. <http://www.bus.utexas.edu/kman/kmprin.
htm>
[13] Dennis, A., Wixom B. & Tegarden Haley (2005). System
Analysis and Design with UML 2.0 2nd Edition. Minion:
Hohn Wiley & Sons Inc.
Aliza Sarlan received her first degree
(Information Technology) from Universiti
Utara Malaysia in 1996 and her Master’s
degree (Information Technology) from
University of Queensland, Australia in 2002.
She is currently lecturing at the Computer
& Information Sciences Department
Universiti Teknologi PETRONAS.
Her
research interests are in Information System
Management and Application Development in Organisational
Context. Currently, her research focuses on the interdependencies
of information technologies and organisational structure, and IS
policy and strategic implementation of ICT in healthcare
industry.
Wan Fatimah Wan Ahmad received her
BA and MA degrees in Mathematics from
California State University, Long Beach,
California USA in 1985 and 1987. She also
obtained a Dip. Ed. from Universiti Sains
Malaysia in 1992. She completed her PhD
in Information System from Universiti
Kebangsaan Malaysia in 2004. She is
currently a senior lecturer at the Computer
& Information Sciences Department of Universiti Teknologi
PETRONAS, Malaysia. She was a lecturer at Universiti Sains
Malaysia , Tronoh before joining UTP. Her main interests are in the
areas of mathematics education, educational technology, human
computer interaction and multimedia.
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
HAND GESTURE RECOGNITION:
SIGN TO VOICE SYSTEM (S2V)
Oi Mean Foong*, Tan Jung Low and Satrio Wibowo
Universiti Teknologi PETRONAS, 31750 Tronoh, Perak Darul Ridzuan, Malaysia
*[email protected]
ABSTRACT
Hand gesture is one of the typical methods used in sign language for non-verbal communication. It is most
commonly used by people who have hearing or speech problems. This paper presents a system prototype
that is able to automatically recognise sign language to help communicate more effectively with the hearing or
speech impaired people. The Sign to Voice system prototype, S2V, was developed using Feed Forward Neural
Network for two-sequence signs detection. The experimental results have shown that the neural network could
achieve a recognition rate of 78.6% for sign-to-voice translation.
Keywords: Hand gesture detection, neural network, sign language, sequence detection.
INTRODUCTION
This system was inspired by the special group of
people who have difficulties to communicate in verbal
form. It was designed with the ease of use for humanmachine interface in mind for the deaf or hearing
impaired people. The objective of this research is to
develop a system prototype that automatically helps
to recognise a two-sequence sign language of the
signer and translate them into voice in real time.
Generally, there are two ways to collect gesture data
for recognition. Device based measurement which
measures hand gestures with equipment such as
data gloves and archive the accurate positions of
hand gestures as its positions are directly measured.
Secondly, vision-based technique which can cover
both the face and hands of the signer where the
signer does not need to wear data glove device. All
processing tasks could be solved by using computer
vision techniques which are more flexible and useful
than the first method [1].
Since the last half of the last century, sign languages
have been accepted as minority languages which
coexist with majority languages [2] and they are the
native languages for many deaf people. The proposed
system prototype was designed to help normal people
communicate with the deaf or mute more effectively.
This paper presents a prototype system known as
Sign to Voice (S2V) which is capable of recognising
hand gestures by transforming digitised images of
hand sign language to voice using the Neural Network
approach.
The rest of the paper is organised as follows: Section
II surveys the previous work on image recognition
of hand gestures. Section III proposes the system
architecture of SV2 prototype. Section IV discusses the
experimental set up and its results, and lastly Section
V draws conclusions and suggests for future work.
This paper was presented at the 5th International Conference in Information Technology
– World Academy of Science, Engineering and Technology, Singapore, 29 August – 2 September 2008
VOLUME Six NUMBER two july - december 2008 PLATFORM
105
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
RELATED WORK
Attempts on machine vision-based sign language
recognition have been published only recently with
relevant literature several years ago. Most attempts
to detect hand gestures/signs from video place
restrictions on the environment. For example, skin
colour is surprisingly uniform so colour-based hand
detection is possible [3]. However, this by itself is not
a reliable modality.
Hands have to be distinguished from other skincoloured objects and these are cases of sufficient
lighting conditions, such as coloured light or gray-level
images. Motion flow information is another modality
that can fill this gap under certain conditions [4], but
for non-stationary cameras this approach becomes
increasingly difficult and less reliable.
Eng-Jon Ong and Bowden [5] presented a novel,
unsupervised approach to train an efficient and
robust detector, applicable not only in detecting the
presence of human hands within an image but also
classifying the hand shape too. Their approach was
to detect the location of the hands using a boosted
cascade of classifiers to detect shape alone in grayscale images.
A database of hand images was clustered into sets of
similar looking images using the k-mediod clustering
algorithm that incorporated a distance metric based
on shape context [5]. A tree of boosted hand detectors
was then formed, consisting of two layers; the top
layer for general hand detection, whilst branches
in the second layer specialise in classifying the sets
of hand shapes resulting from the unsupervised
clustering method.
The Korean Manual Alphabet (KMA) by Chau-Su Lee et
al. [6] presented a vision-based recognition system of
Korean manual alphabet which is a subset of Korean
Sign Language. KMA can recognise skin-coloured
human hands by implementing fuzzy min-max neural
network algorithm using Matrox Genesis imaging
board and PULNIX TMC-7 RGB camera.
106
Image Acquisition (Sign)
Translation (Voice)
Preprocessing
Interpretation
Classification
Image Processing
Object
Neural Network
(Hand Detection)
1.
2.
3.
4.
5.
Sobel Operator
Dilation
XOR Operation
Bounding Box
Proportion Threshold
Figure 1. S2V System Architecture
SYSTEM ARCHITECTURE
Figure 1 shows the system architecture of the proposed
S2V system prototype. Image acquisition for hand
detection is implemented using the image processing
toolbox in MATLAB. This was to develop functions to
capture input from the signer and detect the handregion area. The limitation here is the background
of the image can only be in black color. Therefore
several problems were encountered in capturing and
processing images in RGB values. Thus, an approach
had to be found to detect the hand-region to produce
satisfactory results.
Image Recognition
The input images were captured by a webcam
placed on a table. The system was demonstrated on
a conventional PC Laptop computer running on an
Intel Pentium III Processor with 256 MB of RAM. Each
image has a spatial resolution of 32 x 32 pixels and a
grayscale resolution of 8 bits. The system developed
could process hand gestures at an acceptable
speed. Given a variety of available image processing
techniques and recognition algorithms, it was used as
the design of the preliminary process for detecting the
image as part of image processing. Hand detection
preprocessing workflow is showed in Figure 1.
The system was started by capturing a hand image
from the signer with a webcam, setup towards certain
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
angle with black background. The next process
converted the RGB image into grayscale with either
black (0) or white (1). The edge of each object was
then computed against the black background. The
object was then segmented and differed greatly in
contrast to the background images.
Preprocessing
Changes in contrast was detected by operators
that calculate the gradient of an image. One way
to calculate the gradient of an image is the Sobel
operator [7], [8], [9], which creates a binary mask using
a user-specified threshold value.
The binary gradient mask showed lines of high contrast
in the image. These lines did not quite delineate the
outline of the object of interest. Compared to the
original image, the gaps in the lines surrounding the
object in the gradient mask could be seen.
These linear gaps disappeared if the Sobel image was
dilated using linear structuring elements, applied
by strel function. After finding the holes, the ‘imfill’
function was applied to the image to fill up the
holes. Finally, in order to make the segmented object
look natural, a smoothing process of the object was
applied twice with a diamond structuring element.
The diamond structured element was created by
using the strel function. Then, bitwise XOR operation
set the resulting bit to 1 if the corresponding bit in
the binary image or the result from the dilated image
was a 1. Bitwise manipulation enhanced the wanted
image of the hand region. Further processing with
the bounding box approach was used to clean up the
segmented hand region.
After having a bounding box, the proportion of the
white image as compared to the black image inside
the bounding box was calculated. If the proportion
of white image was changed over the threshold,
then a second image would be captured. Currently,
the prototype used only the two-sequence sign
language.
Figure 2. Feed Forward Neural Network(FFNN)
Classification
Feed Forward Neural network, as shown in Figure 2,
was used in the classification process to recognise the
various hand gestures. It consisted of three layers:
input layer, hidden layer and output layer. Input to the
system included various types of two-sequence sign
language which were converted to a column vector
by neural network toolbox in MATLAB 7.0. The input
signals were propagated in a forward direction on a
layer-by-layer basis. Initialised weight was assigned
to each neuron. The neuron computed the weighted
sum of the input signals and compared the result with
a threshold value, θ . If the net input was less than
the threshold, the neuron output was a value of –1,
otherwise the neuron would become activated and
its output would attain a value of +1 instead. Thus, the
actual output of the neuron with sigmoid activation
function can be represented as
 n

Y = sigmoid  ∑ xi wi − θ 
=
1
i


(1)
In the case of supervised learning, the network was
presented with both the input data and the target
data called the training set. The network was adjusted
based on the comparison of the output and the target
values until the outputs almost matched the targets,
i.e. the error between them is negligible.
However, the dimension of the input for the neural
network was large and highly correlated (redundant).
VOLUME Six NUMBER two july - december 2008 PLATFORM
107
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
EXPERIMENTAL RESULTS
Neural Network Training Technique
Figure 4 shows the comparisons of FFNN training
without PCA and that of Figure 5 FFNN training with
PCA implemented.
The parameter of neural network training are 0.01
in learning rate, epochs in 5000 and momentum
coefficient is 0.8.
Figure 3. Hand Image Detection
It slowed down the processing speed. Thus, the
Principle Component Analysis was used to reduce the
dimension of the input vectors.
In MATLAB, an effective procedure for performing
this operation was the principal component analysis
(PCA). This technique has three effects: the first one
is to orthogonalises the components of the input
vectors (so that they are uncorrelated with each other);
second it orders the resulting orthogonal components
(principal components) so that those with the largest
variation come first; and finally it eliminates those
components that contribute the least to the variation
in the data set. By using this technique, the learning
rate of training the neural network was increased. As
a result, the prototype successfully detected the hand
region as shown in Figure 3.
Figure 4 shows the difference-curve of the error rate
with difference technique before conducting the
training to the neural network. It shows that principle
component analysis increased the learning rate of the
neural network.
The result of the NN training without PCA was MSE
0.050038/0, Gradient 0.114071/1e-010 whereas NN
training with PCA (Fig. 5) was MSE 0.0110818/0,
Gradient 0.00490473/1e-010.
Sequence Detection Testing
Two types of testing were conducted i.e. positive
testing and negative testing. The positive testing was
to prove the sequence of sign language that can be
recognised by the system. The negative testing was
Performance is 0.050038, Goal is 0
0
10
Performance is 0.0110818, Goal is 0
1
10
0
Training-Blue
Training-Blue
10
-1
10
-1
10
-2
10
-2
10
0
500
1000
1500
2000 2500 3000
5000 Epochs
Figure 4. FFNN Training without PCA
108
3500
4000
4500
5000
0
500
1000
1500
2000 2500 3000
5000 Epochs
Figure 5. FFNN Training with PCA
PLATFORM VOLUME Six NUMBER TWO july - december 2008
3500
4000
4500
5000
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
to prove that whenever the sign language was not
moved, the system would not response anything to
the signer. Table I shows the results of these sequence
detection testing.
Table 1. Result Of Sequence Detection Testing
Sequence
1
(+)
Test
(-)
Test
Result
Yes
No
True
Yes
No
True
No
Yes
False
No
Yes
False
Yes
No
True
Recognition Rate
Yes
No
True
70 sets of two-sequence hand gestures were captured
in real-time from signers using a video camera in
which 20 were used as training sets and the remaining
10 were used as test sets. The recognition rate of the
sign languages was calculated as follows:
Yes
No
True
Yes
No
True
The proposed solution was to implement S2V for realtime processing. The system was able to detect the
sequence of sign symbols with additional functions
that was automated to calculate the proportion
of the black and white images and was compared
with a threshold value specified by the program.
The difficulties faced were to recognise a little/small
difference in the portion of the images which were
not detected by the threshold value (and also in
the recognition part). However, for this prototype,
output was obtained by implementing the proposed
technique to detect the sequence .
Recognition rate =
No. of correctly classified signs
× 100 %
Total No. of signs
(2)
The overall results of the system prototype were
tabulated in Table 2 below:
Table 2. S2V System Recognition Rate
Data
No. of
Samples
Recognised
Samples
Recognition
Rate (%)
Training
50
40
80.0
Testing
20
15
75.0
Total
70
55
78.6 (Average)
Sequence
2
The results of segmentation and feature detection
were performed as explained above. Experimental
results of the 70 samples of hand images with different
positions gave consistent outcomes.
Based on the above experiments, the two-sequence
sign language or hand gestures were tested with an
average recognition rate of 78.6%.
CONCLUSION
Hand gestures detection and recognition technique
for international sign languages was proposed and
a neural network was employed as a knowledge
base for sign language. Recognition of the RGB
image and longer dynamic sign sequences was one
of the challenges looked into by this technique.
The experimental results showed that the system
prototype S2V produced satisfactory recognition rate
in the automatic translation of sign language to voice.
For future research, the Hidden Markov Model (HMM)
is proposed to detect longer sequences in large sign
vocabularies and to integrate this technique into a
sign-to-voice system, or vice-versa, to help normal
people communicate more effectively with mute or
hearing impaired people.
VOLUME Six NUMBER two july - december 2008 PLATFORM
109
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
REFERENCES
[1]
Noor Saliza Mohd Salleh, Jamilin Jais, Lucyantie Mazalan,
Roslan Ismail, Salman Yussof, Azhana Ahmad, Adzly
Anuar, and Dzulkifli Mohamad, “Sign Language to Voice
Recognition: Hand Detection Techniques for Vision-Based
Approach,” Current Developments in Technology-Assisted
Education, FORMATEX 2006, vol. 2, pp.967-972.
[2]
C. Neider, J. Kegel, D. MacLaughlin, B. Bahan, and R.G. Lee,
The syntax of American sign language. Cambridge: The MIT
Press, 2000.
[3]
M. J. Jones, and J. M. Rehg, “Statistical Color Models With
Application to skin Detection,” International Journal of
Computer Vision, Jan. 2002, vol. 46, no.1, pp. 81-96.
[4]
D. Saxe, and R. Foulds, “Automatic Face and Gesture
Recognition,” IEEE International Conference on Automatic
Face and Gesture Recognition, Sept. 1996, pp. 379-384.
[5]
E. J. Ong, and R. Bowden, “A Boosted Classifier Tree for Hand
Shape Detection,” Sixth IEEE International Conference on
Automatic Face and Gesture Recognition (FGR 2004), IEEE
Computer Society, 2004, pp. 889-894.
[6]
C. S. Lee, J. S. Kim, G. T. Park, W. Jang, and Z. N. Bien,
“Implementation of Real-time Recognition System for
Continuous Korean Sign Language (KSL) mixed with Korean
Manual Alphabet (KMA),” Journal of the Korea Institute of
Telematics and Electronics, 1998, vol. 35, no.6, pp. 76-87.
[7]
Gonzalez R., and R. Woods, Digital Image Processing.
Addison Wesley, 1992.
[8]
Boyle R., and R. Thomas, Computer Vision: A First Course.
Blackwell Scientific Publications, 1988.
[9]
Davies E., Machine Vision: Theory, Algorithms and
Practicalities. Academic Press, 1990.
O. M. Foong received her BSc Computer
Science from Louisiana State University
and MSc Applied Mathematics from North
Carolina A & T State University in the USA.
She worked as a Computer System Manager
at Consolidated Directories at Burlington,
USA and also as Lecturer at Singapore
Polytechnic prior to joining the CIS
Department at Universiti Teknologi
PETRONAS. Her research interests include data mining, fuzzy
logics, expert systems, and neural networks.
T. J. Low received his BEng (Hons) in
Computer Technology from Teesside
University, UK in 1989 and MSc IT from
National University of Malaysia in 2001.
Low has been in the academic line for the
past 20 years as lecturer in various public
and private institutes of higher learning.
His research interests include wireless
technology, embedded systems, and grid/
HPC computing. Some of his current R&D projects include
Biologically Inspired Self-healing Software, VANET, and Noise
Removal in Seismograph using High Performance Computer. His
other completed projects have been recognised at national as
well as international level, for instance, his Free Space Points
Detector took part in the INPEX 2008 Pittsburgh event in USA in
August 2008. Low has published various papers in systems
survivability, HPC/Grid, and wireless/mobile technology.
[10] X. L. Teng, B. Wu, W. Yu, and C. Q. Liu, “A Hand Gesture
Recognition System Based on Local Linear Embedding,”
Journal of Visual Languages and Computing, vol. 16, Elsevier
Ltd., 2005, pp. 442 – 454.
[11] W. W. Kong, and S. Ranganath, “Signing Exact English
(SEE): Modeling and Recognition,” The Journal of Pattern
Recognition Society, vol. 41, Elsevier Ltd., 2008, pp. 1638
-1652.
[12] Y. H. Lee, and C. Y. Tsai, “Taiwan Sign Language (TSL)
Recognition based on 3D Data and Neural Networks,” Expert
Systems with Applications, Elsevier Ltd., 2007, pp. 1-6.
110
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
ParalleliSation of Prime Number Generation
Using Message Passing Interface
Izzatdin Aziz*, Nazleeni Haron, Low Tan Jung, Wan Rahaya Wan Dagang
Universiti Teknologi Petronas, 31750 Tronoh, Perak Darul Ridzuan, MALAYSIA
*[email protected]
Abstract
This research proposes a parallel processing algorithm that runs on cluster architecture suitable for prime number
generation. The proposed approach is meant to decrease computational cost and accelerate the prime number
generation process. Several experimental results shown demonstrated its viability.
Keywords: Prime number generation, parallel processing, cluster architecture, MPI, cryptography
Introduction
Prime numbers has stimulated much of interest
in the mathematical field or in the security field
due to the prevalence of RSA encryption schemes.
Cryptography often uses large prime numbers
to produce cryptographic keys which are used to
encipher and decipher data. It has been identified
that a computationally large prime number is likely
to be a cryptographically strong prime. However,
as the length of the cryptographic key values
increases, this will result in the increased amount of
computer processing power required to create a new
cryptographic key pair. In particular, the performance
issue is related to time and processing power required
for prime number generation.
Prime number generation comprises of processing
steps in searching for and verifying large prime
numbers for use in cryptographic keys. This is actually
a pertinent problem in public key cryptography
schemes, since increasing the length of the key to
enhance the security level would result in a decrease in
performance of a prime number generation system.
Another trade off resulting from using large prime
numbers pertains to the primality test. Primality test
is the intrinsic part of prime number generation and
is the most computationally intensive sub-process.
It has also been proven that testing the primality of
large candidates is very computationally intensive.
Apart from that, the advent of parallel computing or
processing has invited many interests to apply parallel
algorithms in a number of areas. This is because it has
been proven that parallel processing can substantially
increase processing speed. This paper presents a
parallel processing approach in cluster architecture
for prime number generation that would provide
improved performance in generating cryptographic
keys.
Related Work
Despite the importance of prime number generation
for cryptographic schemes, it is still scarcely
investigated and real life implementations are
of rather poor performance [1]. However, a few
approaches do exist in order to efficiently generate
This paper was presented at the 6th WSEAS International Conference on Computational Intelligence, Man-Machine Systems and Cybernetics, Spain,
14 - 16 December 2008
VOLUME Six NUMBER two july - december 2008 PLATFORM
111
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
prime numbers [1-5]. Maurer proposed an algorithm
to generate provable prime numbers that fulfill
security constraints without increasing the expecting
running time [2]. An improvement was made to
Maurer’s algorithm by Brandt et al. to further speed
up the prime number generation [3]. Apart from
that, the proposed work also included a few ways for
further savings in prime number generation [3]. Joye
et al. presented an efficient prime number generation
scheme that allowed fast implementation on
cryptographic smart card [1]. Besides that, Cheung et
al. originated a scalable architecture to further speed
up the prime number validation process at reduced
hardware cost [4]. All of these research however, were
focused on processing the algorithm sequentially. It
was proved that tasks accomplished through parallel
computation resulted in faster execution as compared
to computational processes that ran sequentially
[9]. Tan et al. designed a parallel pseudo-random
generator using Message Passing Interface (MPI) [5].
This work is almost similar to it but with different
emphasis. The prime numbers generated were used
for Monte Carlo simulations and not cryptography.
Furthermore, considerable progress have been made
in order to develop high-performance asymmetric key
cryptography schemes using approaches such as the
use of high-end computing hardware [6, 7, and 8].
System Model
Experimental Setup
The experimental cluster platform for performance
comprised of 20 SGI machines. Each of the machines
consists of off-the-shelf Intel i386 based dual P3733MHz processors with 512MB memory Silicon
Graphics 330 Visual Workstations. These machines
were connected to a Fast Ethernet 100 Mbps switch.
The head node performs as master node with multiple
network interfaces [10]. Although these machines were
considered to be superseded in terms of hardware
and performance as compared to the latest version of
high performance computers, what was important in
this research was the parallelisation of the algorithm
and how jobs were disseminated among processors.
112
Number Generation
In order to first generate the number, a random seed
was picked and input into the program. The choice
of seed was crucial to the success of this generator
as it had to be as random as possible. Otherwise
anyone who uses the same random function would
be capable of generating the primes, thus defeating
the purpose of having strong primes.
Primality Test
A trial division algorithm was selected as the core for
primality testing. This algorithm was based on a given
composite integer n, trial division consists of trialdividing n by every prime number less than or equal
to n . If a number were found which divides evenly
into n, that number is a factor of n.
Parallel Approach
Once a random number was generated, master node
created a table of dynamic 2D array, which was later
populated with odd numbers. As shown in Figure 1,
a pointer-to-pointer variable **table in master, points
to an array of pointers that subsequently points to a
number of rows. This resulted in a table of dynamic 2D
array. After the table of dynamic 2D array was created,
master then initialised the first row of the table only.
The parallel segment began when master node
broadcasts row [0] to all nodes by using MPI_Bcast.
Row [0] was used by each node to continue populating
Figure 1. Master creates a dynamic 2D array to be populated with
odd numbers
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
the rest of the rows of the table with odd numbers.
Master node then equally divided n-1 number of rows
left that was yet to be populated by number of nodes
available in the grid cluster. Each node was given
an equal number of rows to be populated with odd
numbers. This was achieved by using MPI_Send. A
visual representation of this idea is depicted in Figure
2.
After each node returned the populated rows to
master node, it then picked randomly prime numbers
to be assigned as the value of p, q, and e. These values
were later used for the encryption and decryption
part of a cryptosystem algorithm. It is to be reminded
that the parallel process that took place in the whole
program was only on the prime number generation.
Parallel Algorithm
Each node received n number of rows to be populated
with odd numbers. This was where the parallel
process took place. Each node processed each row
given concurrently. Each node first populated the
rows with odd numbers. Then they filtered out for
prime numbers using the primality test chosen. The
odd prime numbers remained in the rows but those
that were not was assigned to NULL. Each populated
row were then returned to master node, which then
randomly picked three distinct primes for the value of
p,q, and public key e of the cryptographic scheme.
The algorithm of the parallel program is as follows:
Start
Master creates a table of odd numbers and
initialized row [0] only
Master broadcasts row [0] to all slaves
Master sends a number of rows to each slave
Each slave will receive an initialized row
from master
Each slave will populate row prime numbers
Each slave will return populated row to
For an example, if there are 4 processors available
to execute the above tasks, and there are 1200 rows
needed to be populated with prime numbers, each
slave will be given 300 rows to be processed. The
overall procedure is depicted in Figure 3.
Master
Master waits for results from slaves
Master
receives
populated
rows
from
each
slave.
Master checks unpopulated rows
If maxRow > 0
Processor 0 processed row(1) up to row(299),
processor 1 processed row(300) up to row(599),
processor 2 processed row(600) up to row(899) and
lastly processor 3 processed row(900) up to the last
row, row(1199).
Figure 2. Master sends an equal size of row to each slave
Master sends unpopulated row to slave
Master picks prime numbers randomly
Figure 3. Example of assigning 1200 rows to 4 processors (slaves)
VOLUME Six NUMBER two july - december 2008 PLATFORM
113
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
Prompt to select program option
Switch (method)
Case 1: prompt to enter a value greater
than 10000
If value > 10000, generate key primes
Else, Exit program
Case 3: open file and decrypt
Case 4: exit program
End
End
From the figure, it was observed that the algorithm
gather was massive for the first node and deteriorated
as it approached the last node. This may due to the
frequent prime numbers discovered at the beginning
of the number series and and which became scarce
as the numbers became larger towards the end. This
proved that the relative frequency of occurrence of
prime numbers decreases with size of the number
which results in lesser prime numbers being sent back
to master node by later nodes.
Evaluation
Conclusion
Table 1, shows the execution time of running the
parallel program on single and more computing
nodes. From the results, it can be inferred that running
the algorithm in parallel mode has accelerated the
prime number generation process. However, it seems
like there is a noticeable increase in processing time
when the program is running more than 3 nodes. The
execution time recorded was higher when more nodes
participated in the generation process. This may be
caused by network latency during the distribution of
the tasks, which led to the increased of execution time
taken for the communication between nodes.
This study proposed a parallel approach for prime
number generation in order to accelerate the process.
The parallelism and the cluster architecture of this
approach was tested with large prime numbers.
Results demonstrated that improvement can be
obtained if a parallel approach is deployed. However,
further improvements can be made that include:
(1) Use other primality test that is more significant or
feasible for large prime number generation such
as Rabin-Miller algorithm.
(2) Use other random number generation that can
produce random numbers with less computation
yet provides higher security level.
Figure 4 shows the performance measurement using
MPI_GATHER tested on 15 nodes. This figure was
captured using the MPICH Jumpshot4 tool to evaluate
the algorithm usage of MPI libraries. The numbers
plotted shows the amount of time taken for each node
to send back the prime numbers discovered back to
the master node.
114
5
0.043
10
0.053
30
0.093
0 .0 2 47
0.0 25 4
0.02 5 3
0 .0 27 3
0 .0 2 64
0.03 0 3
6
0.015
0 .0 1 37
0.039
0.01 42
3
0.02
0.01
0.005
0
0
7.850
5
0.025
Execution Time (ms)
1
0 .0 27 3
4
0.03
Table 1. Comparison of Execution Time for Different Number of
Nodes.
Number of nodes
0 .0 3 18
3
0.035
0 .0 29 3
0.03 54
0.04
0 .0 32 5
0 .0 3 85
0.045
0.03 5 9
Time (Seconds)
1
2
7
8
9
10 11 12 13 14 15 16
Node
Figure 4. Time taken for MPI_BCAST and MPI_GATHER running
on 15 nodes.
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
References
[1]
M. Joye, P. Paillier and S. Vaudenay, Efficient Generation of
Prime Numbers, Cryptographic Hardware and Embedded
Systems, vol. 1965 of Lecture Notes in Computer Science,
pp. 340-354, Springer-Verlag, 2000
[2]
Maurer, Fast Generation of Prime Numbers and Secure
Public-Key Cryptographic Parameters, Journal of Cryptology,
vol. 8 no. 3 (1995), 123-156
[3]
J. Brandt, I. Damgard, and P. Landrock. Speeding up prime
number generation. In Advances in Cryptology - ASIACRYPT
‘91, vol. 739 of Lecture Notes in Computer Science, pp. 440449, Springer-Verlag, 1991
[4]
Cheung, R. C. C., Brown, A., Luk, W., Cheung, P. Y. K., A Scalable
Hardware Architecture for Prime Number Validation,
IEEE International Conference on Field-Programmable
Technology, 2004. pp. 177-184, 6-8 Dec 2004
[5]
Tan, C. J. and Blais, J. A. PLFG: A Highly Scalable Parallel
Pseudo-random Number Generator for Monte Carlo
Simulations. 8th international Conference on HighPerformance Computing and Networking (May 08 - 10,
2000). Lecture Notes In Computer Science, vol. 1823.
Springer-Verlag, London, 127-135
[6]
Agus Setiawan, David Adiutama, Julius Liman, Akshay
Luther and Rajkumar Buyya, GridCrypt : High Performance
Symmetric Key Cryptography using Enteprise Grids. 5th
International Conference on Parallel and Distributed
Computing, Applications and Technologies (PDCAT 200),
Singapore. Springer Verlag Publications (LNCS Series),
Berlin, Germany. December 8-10, 2004
[7]
Praveen Dongara, T. N. Vijaykumar, Accelerating Privatekey cryptography via Multithreading on Symmetric
Multiprocessors. In Proceedings of the IEEE International
Symposium on Performance Analysis of Systems and
Software (ISPASS), March 2003
[8]
Jerome Burke, John McDonald, Todd Austin, Architectural
Support for Fast Symmetric-Key Cryptography. Proc. ACM
Ninth Int’l Conf. Architectural Support for Programming
Languages and Operating Systems (ASPLOS-IX), Nov. 2000
[9]
Selim G Aki, Stefan D Bruda, Improving A Solution’s Quality
Through Parallel Processing. The Journal of Supercomputing
archive. Volume 19, Issue 2 (June 2001).
Izzatdin Abdul Aziz delves into various
aspects of parallel programming using
Message Passing Interface for High
Performance Computing and Network
Simulation. He worked at PERODUA Sdn
Bhd as a Systems Engineer in 2002.
Presently he is with the Computer and
Information Science Department at
Universiti Teknologi PETRONAS (UTP) as a
lecturer and researcher. He received his Masters of IT from
University of Sydney specialising in Computer Networks.
Nazleeni Samiha Haron is currently
lecturing at the Computer and Information
Science Department, Universiti Teknologi
PETRONAS.
Her research areas are
Distributed Systems as well as Grid
Computing. She received her Masters
Degree from University College London
(UCL) in Computer Networks and
Distributed Systems.
Low Tang Jung is a senior lecturer at
Computer and Information Science
Department,
Universiti
Teknologi
PETRONAS. His research interests are in
High Performance Computing, Wireless
Communications, Embedded Systems,
Robotics and Network Security. He received
his MSc in Information Technology from
Malaysian National University and BEng
(Hons) Computer Technology from Teesside University, UK.
[10] Dani Adhipta, Izzatdin Bin Abdul Aziz, Low Tan Jung, Nazleeni
Binti Haron. Performance Evaluation on Hybrid Cluster:
The Integration of Beowulf and Single System Image, The
2nd Information and Communication Technology Seminar
(ICTS), Jakarta, August 2006
VOLUME Six NUMBER two july - december 2008 PLATFORM
115
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
Evaluation of Lossless Image Compression
for Ultrasound images
Boshara M. Arshin, P. A. Venkatachalam*, Ahmad Fadzil Mohd Hani
Universiti Teknologi PETRONAS, 31750 Tronoh, Perak Darul Ridzuan, Malaysia
*[email protected],
Abstract
Since medical images occupy large amounts of storage space and hinder the speed of tele-healthcare information
exchange, compression of image data is essential. In this work a comprehensive experimental study on lossless
compression algorithms is applied on ultrasound images. There are many compression schemes available for
lossless compression. But in this work, eight commonly used well known algorithms, namely CALIC, JPEG2000,
JPEG-LS, FELICS, Lossless mode of JPEG, SPIHT (with Reversible Wavelets S+P), PNG and BTPC were applied to
determine the compression ratio and processing time. For this study a large number of ultrasound images
pertaining to different types of organs were tested. The results obtained on compression ratio indicated that
CALIC led followed by JPEG-LS. But JPEG-LS took much less time for processing than CALIC. FELICS, BTPC and
the two wavelets-based schemes (SPIHT and JEPG2000) exhibited very close efficiency and they were the next
best to CALIC and JPEG-LS. The results showed that JPEG-LS was the better scheme for lossless compression of
ultrasound images as it demonstrated comparatively better compression efficiency and compression speed.
Keywords: lossless image compression, ultrasound images, compression ratio, compression speed.
INTRODUCTION
Ultrasound is a widely accepted medical imaging
modality for the diagnosis of different diseases due
to its non-invasive, cheap and radiation hazardfree characteristics. As a result of its widespread
applications, huge volumes of data from ultrasound
images have been generated from various medical
systems.
This has caused increased problems
in transmitting the image data to remote places
especially via wireless media for tele-consultation
and storage. Thus efficient image compression
algorithms are needed to reduce file sizes as much as
possible, and make storage access and transmission
facilities more practical and efficient. The relevant
medical information contained in the images must
be well preserved after compression. Any artifact in
the decompressed images may impede diagnostic
conclusions and lead to severe consequences.
Therefore, selecting a suitable method is critical for
ultrasound image coding. This goal is often achieved
by using either lossless compression or diagnostically
lossless compression with moderate compression
ratio. This challenge faced for ultrasound images
motivates the need for research in identifying proper
compression schemes.
There are many approaches to image compression
which can be used. These can be categorised into
two fundamental groups: lossless and lossy. In
lossless compression, the image reconstructed after
decompression is numerically identical to the original
This paper was presented at the International Conference On Man-Machine Systems (ICoMMS 2006) Langkawi, Malaysia,
15-16 September 2006
116
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
image. This is obviously most desirable since no
information is compromised. In lossy compression the
reconstructed image contains degradations relative to
the original to achieve higher compression ratios. The
type of compression group to be used depends on the
image quality needed. Medical images cannot afford
to lose any data. Thus efficient lossless compression
schemes are to be studied.
In the present work, performances of eight efficient
lossless compression schemes namely CALIC,
JPEG2000, JPEG-LS, FELICS, Lossless mode of JPEG,
SPIHT with Reversible Wavelets (S+P), PNG and
BTPC were studied to investigate the compression
efficiency for digital ultrasound images highlighting
the limitation of cutting edge technology in lossless
compression. The study focused mainly on the
effectiveness of compression and computational time
of each compression method on ultrasound images.
statistical modeler and coder. The modeler gathers
information about the data by looking for some
context and identifies a probability distribution that
the coder uses to encode the next pixel xi+1 It can be
viewed as following an inductive statistical inference
problem where an image is observed in raster-scan.
At each instant i, after having observed the subset of
past source sample xi = (x1, x2, … , xi), but before
observing xt+1 a conditional probability distribution
P(.|xi) is assigned to the next symbol xi+1 Ideally, the
code length l contributed by xt+1 is
l (xi+1) = – log P(xi+1|xi) bits,
which averages to the entropy of the probabilistic
model. Modern lossless image compression [2, 3] is
based on the above paradigm in which the probability
assignment is broken into the following steps:
•
LOSSLESS IMAGE COMPRESSION
A typical image compression implementation consists
of two separate components, an encoder and a
decoder as shown in Figure 1. The former compresses
the original image into a more compact format
suitable for transmission and storage. The decoder
decompresses the image and reconstructs it to the
original form. This process will either be lossless or
lossy, which will be determined by the particular
needs of the application.
Most of the recent efficient lossless compressions
can be classified into one of the three paradigms:
predictive with statistical modeling, transform-based
and dictionary-based schemes. The first paradigm
typically consists of two distinct components:
Input
Image
Encoder
Data Storage
Or
Transmission
Decoder
Figure 1. A typical data compression system
Output
Image
•
•
A predictive step, with the advantage of image
information redundancy (correlation of data) to
for
construct an estimated prediction value
the next sample xi+1 based on a finite subset of
the available past source symbols xi .
The determination of context in which a value xi+1
occurs. The context is a function of a (possible
different) causal template.
A probabilistic model for prediction error ei +1 = x i +1 - x̂ i +1 , conditioned on the context of
xi+1
The prediction error is coded based on a probability
distribution using a simple function of previously
neighbouring samples (W, N, E and NW) as shown in
Figure 2. The advantage of prediction is that it decorrelates the data samples thus allowing the use of a
simple model for the prediction errors [2].
NW N
W ?
E
Figure 2. Four neighbouring samples in a causal template.
VOLUME Six NUMBER two july - december 2008 PLATFORM
117
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
In transform-based coding, the original digital image
is transformed into frequency or wavelet domain
prior to modeling and coding, where it is highly decorrelated by the transform. This de-correlation
concentrates the important image information into
a more compact form in which the redundancy can
be removed using an entropy coder. JPEG2000 is a
transform-based image compression scheme.
Finally, the dictionary-based compression algorithms
substitute shorter codes for longer patterns of strings
within the data stream. Pixel patterns (substrings) in
the data stream found in the dictionary are replaced
with a single code. If a substring is not found in the
dictionary, a new code is created and added to the
dictionary. Compression is achieved when smaller
codes are substituted for longer patterns of data, such
as GIF and PNG which are widely used in the Internet.
In this work focused on the following lossless
compression schemes:
1. CALIC which combines non-linear prediction with
advanced statistical error modeling techniques.
2. FELICS in which each pixel is coded in the context
of two nearest neighbours.
3. BTPC, a multi-resolution technique which
decomposes the image into a binary tree. These
methods are the result of improvements in
prediction and modeling techniques, as well as
improvements in coding techniques that yielded
significantly higher compression ratios [5].
4. Lossless mode of JPEG which combines simple
linear prediction with Huffman coding.
5. JPEG-LS is based on Low Complexity Lossless
Compression method LOCO-I, which combines
good performance with efficient implementation
that employs nonlinear simple edge detector
prediction, in particular with Golomb-Rice coding
or Run Length Coding.
118
6. Lossless mode of JPEG2000 which is integerto-integer wavelets based on Embedded Block
Coding with block Truncation (EBCOT) [5].
7. SPIHT with Reversible Wavelets (S+P) as a
second lossless wavelet-based method beside
JPEG2000.
8. Portable Network Graphics (PNG) which is a
dictionary-based compression method [1, 2, 3].
Artur Przelaskowski [6] considered only four effective
lossless compression schemes: Binary context-based
arithmetic coder, CALIC, JPEG-LS and JPEG2000 for 22
mammogram images. This study pioneered different
compression methods for mammograms; however
only a few lossless coders were involved and it was
found that the best lossless compression that could
be achieved was only 2.
Kivijärvi, et al. [7, 9] studied lossless compression of
medical images of various modalities: Computed
Radiography, Computed Tomography, Magnetic
Resonance Imaging, Nuclear Medicine, and
Ultrasound. Performance criteria of a range of general
and specific image lossless compression schemes
were evaluated. They observed that CALIC performed
consistently well and achieved compression ratio
2.98. JPEG-LS compression ratio was almost as well
at 2.81 and lossless JPEG with Huffman encoding
and selection prediction value five [1] had a lower
compression ratio 2.18. PNG did not perform well as it
only achieved compression ratio 1.90, and neither did
any of the general-purpose compression schemes.
These early studies did not have the opportunity to
examine the performance of the JPEG2000 scheme.
The measurement of compression ratio in their study
was the average compression ratios of all modalities,
regardless of the fact that different modalities may
appear to have different redundancy which would
lead to differences in compression ratios. They also
measured compression and decompression time, and
concluded that “CALIC gives high compression in a
reasonable time, whereas JPEG-LS is nearly as effective
and very fast”.
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
Denecker [8] used more than five image based and
general-purpose compression schemes, including
lossless JPEG, CALIC, S+P and GNU UNIX utility gzip.
They were tested on Computed Tomography, MRI,
Positron Emission Tomography, Ultrasound, X-Ray
and Angiography images. It was indicated that CALIC
performed best with an average compression ratio
of 3.65. S+P with arithmetic encoding achieved 3.4,
lossless JPEG with Huffman encoding achieved about
2.85 and gzip achieved about 2.05. Since it was not
stated as to how many images were used in this study,
it is difficult to interpret the average compression
ratios specified [10].
METHODOLOGY
The different lossless Compression algorithms referred
above yielded different values for compression ratio
and speed for the same type of images. It is essential
to decide which one among them will be better
when applied on an image. Two most important
characteristics of lossless image compression
algorithms are Compression Ratio (CR) and
Compression Speed (CS). If the combined time taken
for compression and decompression is small then
CS is high. These characteristics help to determine
the suitability of a given compression algorithm to
a specific application. CR is simply the size of the
original image divided by the size of the compressed
image. This ratio gives an indication of how much
compression is achieved for a particular image.
In order to calculate how much time the process
(compression or decompression) has used in seconds,
the function clock ( ) in C++ was used. This function
was called at the start and at the finish of the process
to compute the duration in seconds (as shown in the
following equation) using the CPU ticks since the
system was started as follows:
decompression time, some compression methods are
symmetric which means that equal time is taken for
compression and also decompression. An asymmetric
algorithm takes more time to compress than to decompress. All the compression algorithms used in
this study were implemented using MS Visual C++
along with JASPER, MG, LIBPNG, HP implementation
of JPEG-LS and ZLIB. For this study, a large number
of ultrasound images which were different in texture
and size were used. The results obtained for a sample
set of 21 images applying the above algorithms
for compression efficiency were analysed. The
ultrasound images were obtained from local hospitals
and scanned in the Intelligent Imaging Laboratory at
Universiti Teknologi PETRONAS using a Pentium IV
processor with 2.8 MHz under Windows XP.
RESULTS and DISCUSSIONS
Compression Efficiency
The results obtained for a test sample set of 21 images
applying the above method for compression efficiency
are shown in Figure 3. It was observed that the two
predictive state-of-the-art CALIC and JPEG-LS were
the first two efficient schemes for ultrasound images
independent of texture and size.
Although the best achievable compression ratio
chosen was with LJPEG having tried all the possible
seven predictors for lossless coding [1], LJPEG was
found to be lagging behind all schemes in terms of
efficiency. LJPEG used a very simple linear prediction
that combined three surrounding neighbouring
Duration = (finish – start) / CPU ticks per second
This duration value depended on the complexity
and efficiency of the software implementation of
the compression/decompression algorithms and the
speed of the processor. Considering compression/
Figure 3. Compression efficiency
VOLUME Six NUMBER two july - december 2008 PLATFORM
119
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
pixels. PNG which is a dictionary-based scheme
outperforms LJPEG only slightly. FELICS, BTPC and
the two wavelets-based schemes (JPEG2000 and
S+P) were found to lie in between and performing
almost equally good with slight differences in terms of
efficiency after CALIC and JPEG-LS. From the results,
it clearly depicted that predictive schemes with
statistical modeling performed better than transform
based coding which outperformed the dictionary
based coding for lossless compression of images.
Compression Speed
Figures 4(a) and (b) show the compression and
decompression times respectively for the above
eight lossless schemes. It can be seen that the FELICS
was the fastest and CALIC seemed to be extremely
slow, whereas LJPEG and JPEG-LS were somewhat
much closer to FELICS. It should be noted that the
above four algorithms are symmetric. PNG and BTPC
demonstrated less compression speeds compared to
the two wavelet-based algorithms.
(a) Compression time in Seconds
From the above analysis, it was found that the
predictive schemes yielded improved compression
efficiency and speed, and resulted in reduced cost of
transmission and storage especially for tele-healthcare
applications. JPEG-LS with embedded run length
encoding was found to be better than other methods
for lossless compression of ultrasound images.
CONCLUSION
In this paper a comparative study on lossless
compression schemes applied on ultrasound images
was carried out. To evaluate the lossless compression
methods, three criteria were studied viz., compression
ratio, compression time and decompression time. A
large number of ultrasound images were processed
but only the results for a set of 21 test cases were
included (refer Table 1). Various aspects of the lossless
compression procedures were examined. CALIC
gave the best compression efficiency but much
lower compression speed. JPEG-LS showed high
compression ratio (very close to CALIC) and better
compression speed than CALIC. Based on these two
features, the results showed that JPEG-LS is well
suited for compressing ultrasound images. Additional
research is to be carried by applying JPEG-LS on the
regions of interest and lossy compression algorithms
on the remaining portion of the ultrasound image.
This will will not only result in efficient storage but will
also increase speed of transmission of images in telehealthcare applications.
ACKNOWLEDGEMENT
The authors would like to thank the Department of Diagnostic
Imaging of Ipoh Hospital for providing the medical images used
in this work.
REFERENCES
[1]
Khalid Sayood, Introduction to Data Compression, Morgan
Kaufmann Publishers, 2000
[2]
B. Carpentieri, et al., Lossless compression of continuoustone images, Proc. IEEE 88 (11) (2000) I797-1809
[3]
Randers-Pehrson G, et al., PNG (Portable Network Graphics)
specification version 1.1. PNG Development Group.
February 1999
(b) Decompression time in Seconds
Figure 4. Compression/decompression time
120
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
[4]
John A. Robinson, “Efficient general-purpose image
compression with binary tree predictive coding”, IEEE
Transactions on Image Processing, Apr 1997
[7]
Kivijärvi J, et al., “A comparison of lossless compression
.methods for medical images”, Computerized Medical
Imaging and Graphics, 1998
[5]
D. Taubman,“High performance scalable image compression
with EBCOT”, IEEE Trans. On Image Processing, July 2000
[8]
[6]
Artur Przelaskowski, ”Compression of mammograms for
medical practice”, ACM symposium on Applied Computing,
2004
Denecker K, Van Overloop J, Lemahieu I, “An experimental
comparison of several lossless image coders for medical
images”, Proc. 1997 IEEE Data Compression Conference,
1997
[9]
David A. Clunie, ”Lossless Compression of Grayscale
Medical images: Effective of Traditional and State of the Art
approaches”.
Table 1. Compression ratios, compression/decompression time
Images
Ultrasound1
Ultrasound2
Ultrasound3
Ultrasound4
Ultrasound5
Ultrasound6
Ultrasound7
Ultrasound8
Ultrasound9
Ultrasound10
Ultrasound11
Ultrasound12
Ultrasound13
Ultrasound14
Ultrasound15
Ultrasound16
Ultrasound17
Ultrasound18
Ultrasound19
Ultrasound20
Ultrasound21
CR
CT DT
CR
CT DT
CR
CT DT
CR
CT DT
CR
CT DT
CR
CT DT
CR
CT DT
CR
CT DT
CR
CT DT
CR
CT DT
CR
CT DT
CR
CT DT
CR
CT DT
CR
CT DT
CR
CT DT
CR
CT DT
CR
CT DT
CR
CT DT
CR
CT DT
CR
CT DT
CR
CT DT
LJPEG
2.078
0.203 0.094
1.828
0.188 0.078
1.897
0.25 0.078
1.816
0.187 0.078
1.904
0.187 0.078
1.815
0.203 0.078
2.075
0.375 0.078
2.078
0.203 0.14
2.045
0.203 0.093
1.670
0.203 0.078
1.893
0.203 0.078
1.788
0.203 0.078
1.867
0.203 0.078
1.794
0.203 0.078
1.879
0.187 0.078
2.139
0.187 0.063
2.024
0.25 0.078
1.748
0.203 0.078
2.197
0.203 0.093
2.175
0.218 0.078
2.167
0.187 0.062
JPEG-LS
2.302
0.234 0.235
2.074
0.203 0.203
2.188
0.265 0.203
2.072
0.234 0.203
2.190
0.203 0.203
2.032
0.218 0.406
2.326
0.375 0.219
2.302
0.234 0.235
2.297
0.234 0.235
1.931
0.235 0.235
2.217
0.234 0.235
2.015
0.219 0.265
2.193
0.218 0.218
2.054
0.203 0.219
2.195
0.218 0.219
2.507
0.203 0.203
2.382
0.187 0.187
1.937
0.234 0.234
2.455
0.234 0.25
2.447
0.234 0.234
2.516
0.219 0.219
JPEG2000
2.194
1.156 0.937
1.952
1.125 0.921
2.028
1.109 0.937
1.946
1.14 0.968
2.028
1.14 1.015
1.934
1.171 1.015
2.172
1.312 1.109
2.194
1.187 0.906
2.166
1.234 1.14
1.840
1.25 1.046
2.063
1.312 1.046
1.924
1.187 1.156
2.043
1.156 1.14
1.929
1.234 1.171
2.039
1.328 1.125
2.375
1.078 0.906
2.117
1.125
1
1.877
1.171 1.14
2.374
1.187 1.109
2.377
1.234 1.14
2.389
1.203 1.031
CALIC
2.381
2.703 2.703
2.128
2.5
2.5
2.242
2.468 2.468
2.133
2.5
2.5
2.244
2.484 2.468
2.100
2.531 2.531
2.412
2.718 2.718
2.381
2.781 2.765
2.380
2.775 2.765
1.991
2.625 2.562
2.254
2.515 2.515
2.069
2.609 2.593
2.223
2.562 2.515
2.096
2.546 2.546
2.232
2.531 2.531
2.575
2.375 2.375
2.433
2.359 2.328
2.012
2.593 2.546
2.519
2.812 2.812
2.525
2.843 2.843
2.582
2.421 2.593
FELICS
2.250
0.127 0.113
1.980
0.141 0.122
2.087
0.126 0.109
1.978
0.127 0.112
2.093
0.126 0.111
1.964
0.13 0.116
2.263
0.128 0.12
2.250
0.127 0.116
2.234
0.139 0.127
1.821
0.147 0.121
2.094
0.124 0.114
1.927
0.136 0.117
2.034
0.124 0.111
1.952
0.127 0.116
2.070
0.143 0.113
2.324
0.116 0.106
2.274
0.119 0.106
1.864
0.133 0.122
2.359
0.128 0.115
2.336
0.133 0.125
2.349
0.118 0.109
PNG
2.133
0.64 0.263
1.923
0.546 0.256
1.992
0.5 0.371
1.912
0.515 0.252
1.999
0.609 0.253
1.904
0.5
0.25
2.133
0.765 0.294
2.133
0.765 0.287
2.113
0.625 0.27
1.816
0.625 0.274
2.016
0.656 0.553
1.898
0.484 0.252
2.000
0.671 0.253
1.902
0.5 0.251
2.002
0.671 0.25
2.188
0.609 0.244
2.120
0.515 0.25
1.865
0.593 0.248
2.247
0.765 0.266
2.248
0.75 0.267
2.201
0.765 0.243
BTPC
2.216128805
0.718 0.406
1.930
0.75 0.375
2.033
0.75 0.453
1.932
0.796 0.375
2.033
0.703 0.5
1.918
0.796 0.375
2.230
0.812 0.39
2.216
0.843 0.64
2.198
0.875 0.625
1.780
0.859 0.39
2.042
0.703 0.703
1.879
0.859 0.39
1.993
0.828 0.406
1.900
0.859 0.375
2.019
0.843 0.359
2.326
0.656 0.609
2.213
0.812 0.359
1.824
0.859 0.39
2.349
0.843 0.375
2.325
0.875 0.406
2.348
0.843 0.343
S+P
2.208
0.858 0.612
1.945
0.936 0.723
2.005
0.889 0.583
1.950
0.917 0.706
2.0127
0.91 0.711
1.938
0.945 0.801
2.183
0.999 0.584
2.208
0.962 0.864
2.185
1.002 0.84
1.838
1.074 0.917
2.057
0.972 0.836
1.922
1.034 0.65
2.026
0.99 0.885
1.919
1.01 0.904
2.032
0.997 0.849
2.390
0.836 0.721
2.0886
0.976 0.574
1.866
1.037 0.954
2.397
0.917 0.79
2.383
0.942 0.778
2.429
0.801 0.73
VOLUME Six NUMBER two july - december 2008 PLATFORM
121
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
Learning Style Inventory System:
A Study To Improve Learning Programming Subject
Saipunidzam Mahamad*, Syarifah Bahiyah Rahayu Syed Mansor, Hasiah Mohamed @ Omar1
*Universiti Teknologi Petronas, Bandar Seri Iskandar, 31750 Tronoh Perak, Malaysia
1UiTM Cawangan Terengganu, 23000 Dungun, Terengganu, Malaysia
*[email protected]
ABSTRACT
This paper presents a learning style for personal development as an alternative for a student in learning a
programming subject. The model was developed with the intention to increase awareness of learning style
preferences among higher education students and as a guide to excel in the programming subject. The aim
of this paper is to integrate the ideal learning style with individual learning preferences in improving the way
of learning a programming subject. The study proposes a learning style model based on a human approach
to perform a task, and human emotional responses to enable learning to be oriented according to a preferred
method. The learning style defines the basic learning preference to make the assessment. A prototype, Learning
Style Inventory System (LSIS) was developed to discover an individual learning style and recommend how
to manipulate the learning skill. The analysis and comparisons showed that personal learning styles play an
important role in a learning process.
Keywords: Learning Style Inventory System, Learning Style, Learning Style Model.
INTRODUCTION
Learning a programming subject seems to be difficult.
Many students claim to dislike the programming
subject which may have led to their inability to do
programming. The main reason could be that students
do not understand or know how to manipulate their
learning style. Knowing what we have (strengths and
weaknesses) and how to improve learning is crucial.
There is a need to have a system with capabilities
of discovering one’s own learning style, and how to
improve it in order to assist the students, especially
those facing a transition period from secondary
school to university environment. Once students are
actively engaged in their own learning process, they
will begin to feel empowered and indirectly, develop
their personal sense of achievement and self-direction
levels. Thus, it may improve the student’s performance
in their future undertakings.
The key to get and keep students actively involved
in learning lies in understanding learning style
preferences, which can positively or negatively
influence a student’s performance (Hartman, 1995).
Learning style enables learning to be oriented
according to the preferred method. Besides, a study
showed that adjusting teaching materials to meet
the needs of a variety of learning styles benefits the
students (Kramer-Koehler, Tooney & Beke, 1995).
Therefore, a person’s learning style preferences have
an impact upon his/her performance in learning.
This paper was presented at the Seminar It Di Malaysia 2006 (SITMA06),
19 - 20 August 2006
122
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
In this paper, the authors analyse correlations between
a learning style with the student’s performance
in learning a programming subject. To further
investigate the correlation, a system named Learning
Style Inventory System (LSIS) was developed. This
system was built based on Kolb’s learning style model
and a strategies shift learning model by Richard M.
Felder and Barbara A. Solomon. The strategies to
shift learning model is used to recommend students
on how to improve his/her learning style. This
recommendation feature consisted of strategies of
moving between two-continuum approaches in
developing new learning style preferences focusing
on learning a programming subject. This study also
observed the implications of learning style towards
learning a programming subject.
LEARNING STYLE MODEL
Nowadays, there are quite a number of learning style
preferences being introduced. Some of them are
Kolb’s Model (David A. Kolb, 2001), Multiple Intelligent
(David Lazear, 1991), Vision Auditory and Kinesthetic
model (Dawna Markova, 1996). Each of the models
has its own strengths and weaknesses.
Kolb’s learning styles model is based on two lines of
axis; (1) human approach to a task that prefers to do
or watch; and (2) human emotional response which
prefers to think or feel, as shown in Figure 1. The
east-west axis is the Processing Continuum, while the
north-south axis is the Perception Continuum.
The theory consists of four preferences, which are the
possible different learning methods:
Concrete experience
Active
experimentation
(4)
(1)
DYNAMIC
LEARNER
INNOVATIVE
LEARNER
(3)
(2)
COMMON
SENSE
LEARNER
ANALYTIC
LEARNER
Abstract conceptualization
Figure 1. Learning style type grid
Reflective
observation
Doing Watching Feeling Thinking –
–
–
–
active experimentation
reflective observation
concrete experience
abstract conceptualization
According to David A. Kolb (2001), the learning cycle
begins at any one of the four points where the
approach should be in a continuous spiral mode.
However, it was suggested that the learning process
often begins when a person performs a particular
action. Later, he/she will discover the effect of the
action in that situation. Lynda Thomas et al. (2002)
agreed with Kolb. She indicated that the process
should be in the sequence order, where the second
step follows the first step. In the second step, a person
should understand these effects in that particular
instance. In case if the same action were undertaken
in the same circumstances, it would be possible to
anticipate what would follow from the action. Then,
in this pattern, the third step would understand the
general principle under which the particular instance
occurred.
According to James Anderson (2001), to modify a
person’s learning style is necessary, and it is needed
in a particular situation which forces them to enhance
their capability to move along with the two continuums
as discussed in Kolb’s model. He stated that a person
himself is responsible to change his/her learning style
to meet the requirements of a particular situation.
Due to lack of knowledge, a student usually tended to
be in the current learning style preference, and would
be rigid to move along between the two continuums.
The cause might be that they have no idea what is
their learning style, and how to improve it. A study
by Richard M. Felder and Barbara A. Solomon (2000)
illustrated strategies to shift a student’s learning style
such as active experimentation, reflective observation,
concrete experience and abstract conceptualisation
as shown in Table 1.
THE LEARNING STYLE INVENTORY SYSTEM (LSIS)
The LSIS, a standalone system, consisted of several
forms to indicate different levels of process flow. The
aim of this model was to discover how students learn
VOLUME Six NUMBER two july - december 2008 PLATFORM
123
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
Table 1. Strategies to shift a learning style
Learning style
Problem
Suggestion
Active learner
Stuck with little or no class
discussion or problem-solving
activities
Group discussion or Q&A sessions
Reflective
learner
Stuck with little or no class
time for thinking about new
information
Write short summaries of readings or class notes in own
words.
Conceptual
learner
Memorization and substitution
formula
Find link/connection on the facts with the interpretation
or theories
Experience
learner
Material is abstract and
theoretical
Find specific example of concept and procedure and
find out how the concepts apply in practice by asking
instructor, or referring to some course text or other
references or brainstorming with friends or classmates
without the intention of evaluating their learning
ability. The objectives of the system were to help
students discover their learning style type and give
recommendations towards each type of learning
to maximise their learning skills upon studying a
programming subject, in particular. The overall
system flow is shown in Figure 2.
The most crucial phase is the completion of the learning
style test. In this phase, the system accumulates
learning result to determine a student’s learning style,
hence it gives a recommendation on how to develop
his/her learning style towards learning a programming
subject. The system is divided into three categories.
Category 1 – Develop Learning Style (LS) inventory to
obtain students learning style using multiple-choice
tests, which implemented based-on the Kolb’s scale.
The learning style inventory describes the way a
person learns and how to deal with ideas and day-today situations in life.
Category 2 – Calculate and analyse the test answer
and produce results of the student learning style with
a description.
Category 3 – Propose an appropriate recommendation
to student’s original learning type to suit with the
requirement on learning a programming subject.
124
Figure 2. System flow
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
26%
74%
Aware
Unaware
Figure 4. Awareness of
learning style preferences
41%
46%
13%
Aware
Claim Aware but eventually Unaware
No
Figure 5. Knowledge of own
learning style type
Figure 3. Screen shot for the LSIS result form interface
The sample screen shot of the result page is shown
in Figure 3. The page provided students with their
personal learning style as well as its description. Also,
it displayed some recommendation and learning
methods.
THE EXPERIMENT
The experiment was conducted during the July 2004
semester at Universiti Teknologi PETRONAS (UTP),
Malaysia. The total number of participants was 106
students. Those were students from foundation and
first year undergraduate Information Communication
and Technology (ICT) and Business Information System
(BIS) programmes. At that time, these students were
taking an introductory programming course, and the
course was in a series of programming courses offered
to ICT and BIS students.
Two sets of questionnaires were given to the students.
Both sets were designed to obtain information about
the students’ academic performance, and to know
their views on learning styles and the programming
subject. The first set was distributed during the
beginning of the semester together with LSIS, which
asked the students about the knowledge in learning
styles and experiences on learning programming. The
LSIS was used to identify the students’ learning style
and to suggest a better way to learn the programming
subject. The second set was distributed during the
last class session of the course to analyse the students’
performance in learning programming, and how the
personal learning style played an important role in
their learning process.
Below were the steps taken in analysing the data from
both sets of questionnaires:
• The questions were divided into six categories
in order to capture the students’ knowledge of
learning style, knowledge of own learning style
type, tabular of learning style type, liking toward
the programming subject, level of difficulty faced
and learning style implications toward learning
programming.
VOLUME Six NUMBER two july - december 2008 PLATFORM
125
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
•
•
The percentages of answers for each category
were counted.
The students’ learning style for the first set of
questionnaires was compared to the second set
to determine the positive result of LSIS and its
correlation with the achievement in the course.
50
35
39
Analytic
Common Sense
40
30
20
18
14
Dynamic
Innovative
10
0
RESULTS AND DISCUSSIONS
Figure 6. Tabular for learning style type
This section is divided into two parts. The first part
illustrates the participants’ awareness of their learning
style using a questionnaire and LSIS, while the other
part of the section was to discover the implications
of their learning style type towards learning
programming.
The first category of questionnaire was to capture the
students’ awareness of the existence of their Learning
Style Preferences. Figure 4 shows that only 26% of
the respondents had knowledge of a Learning Style
Model.
The second category of questions, shown in Figure
5, was designed to know the awareness of students
of their own self-learning style type. The results
showed that 49 respondents had no idea about their
learning style while 57 respondents said that they
do know about it. However, after using LSIS only 14
respondents had the correct learning as claimed.
The table for learning style type as shown in Figure
6 illustrates the number of persons for each learning
style type. It consists of four categories, which are
dynamic, innovative, analytic and common sense
learner for both samples. The respondents were using
LSIS to discover their learning style. The collected
total number for each type of learner represents
the percentage of students as a whole. The analysis
may help lecturers apply the most suitable way of
delivering lectures to suit the students’ learning
styles. Moreover, the students could follow the given
recommendations to improve their way of learning
the programming subject.
Table 2 indicates the students’ mentality state of liking
the programming subject. The graph clearly shows
that 83 (79%) respondents liked the programming
subject compared to respondents who disliked
programming. This is a positive feedback to ease
the lecturer in tackling students as most students
showed a positive attitude towards programming
and would be willing to change the way they learn as
recommended by LSIS.
The level of difficulty faced by a student while learning
programming is shown in Table 3. Evaluation from
the statistic shows that up to 56% (equivalent to 59
respondents) think that learning a programming
subject is challenging and that they have to study
hard to excel in the programming subject.
The results shown in Table 2 and Table 3 were
for students who had the experience in learning
programming before they undertook the current
subject. This valuable information could have
motivated the students to accept the recommendation
proposed by LSIS. However, the students could choose
or not choose to follow the LSIS recommendation.
Table 2. Statistic on Likeness towards Programming Subject
A
A–
B+
B
C+
C
Yes
14
7
8
7
3
No
2
1
2
2
Likeness
126
Taking
TOTAL
3
42
83
1
9
23
PLATFORM VOLUME Six NUMBER TWO july - december 2008
D+
D
F
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
Table 3. Statistic on Level of Difficulties to Learn Programming Subject
Level
A
A–
B+
B
C+
Very easy
2
Easy
1
2
Medium
6
4
1
2
1
Challenging
5
2
7
7
4
Hard
2
C
D+
D
F
Taking
TOTAL
2
4
3
6
3
17
33
1
26
59
2
4
Table 4. Implication of learning style type toward learning programming subject
Learning Style Type
A
A–
B+
B
Dynamic
3
2
8
4
Innovative
8
6
2
1
3
20
Analytic
6
3
5
3
1
18
Common Sense
1
4
4
7
1
Learning style implications toward learning a
programming subject shown in Table 4. The purpose
of this analysis was to observe students’ action on
the recommendation in order to improve the way of
learning the programming subject throughout the
semester. From the analysis, there were 74 respondents
who followed the recommendation and take that as a
part of their way of learning. Later, the result was used
to compare the with student’s coursework mark to
analyse their achievement.
Overall, the results of the course were very interesting.
The experiment involved the students’ initiative to
change and follow the recommendations in learning a
programming subject. Besides, the environment was
designed where students had control of their learning
experience. This study discovered that learning style
has some impact on learning.
CONCLUSION
The LSIS was successfully developed with the aim
of increassing the awareness of Learning Style
Preferences among students, and guide them to
excel in a programming language subject. LSIS is a
C+
C
D+
D
F
TOTAL
17
2
19
computer system to assist students in identifying
and analysing learning styles based on Kolb’s model
and the strategies shift learning model by Richard
M. Felder and Barbara A. Solomon, and then guide
students to improve their learning style perspective.
Preliminary evaluation indicated that students lacked
knowledge in learning style preference and their
own learning style type. The implementation of the
system benefited students in cope with their daily
learning routine as well as increase their confidence
level in learning a programming subject.
For future work, researchers may want to integrate
other types of learning styles and features of advising
students. It could include a systematic shifting
procedure between the best suitable learning style
towards learning a programming subject. Therefore,
users could discover their learning styles from the
various models and improve the learning style on
learning a programming subject.
ACKNOWLEDGEMENT
The authors would like to thank Aritha Shadila Sharuddin for
conducting the survey for data gathering.
VOLUME Six NUMBER two july - december 2008 PLATFORM
127
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
REFERENCES
[1]
David A. Kolb (2001). Experiential Learning Theory
Bibliography 1971-2001. http://trgmcber.haygroup.com/
Products/learning/bibliography.htm
[2]
David Lazear (1991). Seven Ways of Knowing: Teaching for
Multiple Intelligences. Skylight Publishing.
[3]
Dawna Markova (1996). VAK Learning Style. http://www.
weac.org/kids/june96/styles.htm
[4]
James Anderson (2001). Tailoring Assessment to Student
Learning Styles http://aahebulletin.com/public/archive/
styles.asp
[5]
Kramer-Koehler, Pamela, Nancy M. Tooney, and Devendra
P. Beke (1995). The Use of learning style innovations to
improve retention. http://fairway.ecn.purdue.edu/asee/
fie95/4a2/4a22/4a22.htm
[6]
Lynda Thomas et al. (2002). Learning Style and Performance
in the Introductory Programming Sequence. ACM
1-58113.473-8/02/0002.
[7]
Richard M. Felder & Barbara A. Solomon (2002). A
Comprehensive Orientation and Study Skills Course
Designed for Tennessee Families First Adult Education
Classes.
[8]
Virginia F. Hartman (1995). Teaching and learning style
preferences: Transitions through technology. http://www.
so.cc.va.us/vcca/hart1.htm
128
Saipunidzam Mahamad was born in
Perak, MALAYSIA on October 23, 1976. The
author was awarded the Diploma in
Computer Science on 1997 and the
Bachelor Degree in computer science in
1999 from University Technology of
Malaysia, Johor. Later, he pursued his
Master’s
degree in Computer and
Information Science at University of South
Australia and was awarded a Master’s Degree in 2001. He is a
lecturer since January 2002 and is currently attached with
University Technology PETRONAS. He started with University
Technology PETRONAS (UTP) as a trainee lecturer since July 2000.
Apart from teaching, he is involved in Software Engineering and
the E-Commerce research group and does various administrative
work. Recently, he was appointed as UTP’s e-learning manager to
manage and monitor the implementation of e-learning at UTP.
His research interests include Computer System, E-Learning,
Software engineering and Intelligent Systems.
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
Performance Measurement
– A Balanced Score Card Approach
P. D. D. Dominica*, M. Punniyamoorthy1, Savita K. S. and Noreen I. A.
*Universiti Teknologi PETRONAS, 31750 Tronoh, Perak
1National Institute of Technology, Tiruchirapalli, INDIA
*[email protected]
ABSTRACT
This paper suggests a framework of performance measurement through a balanced scorecard and to provide
an objective indicator for evaluating the achievement of the strategic goals of the corporate. This paper
uses the concepts of a balanced score card and adopts an analytical hierarchical process model to measure
an organisational performance. The balanced score card is a widely used management framework for the
measurement of organisational performance. Preference theory is used to calculate the relative weightage for
each factor, using the pair wise comparison. This framework may be used to calculate the effectiveness score
for balanced score card as a final value of performance for any organisation. The variations between targeted
performance and actual performance were analyzed.
Keywords: Balanced score card, corporate strategy, performance measurement, financial factors and non-financial
factors.
INTRODUCTION
Over the recent past, organisations have tried various
methods to create an organisation that is healthy
and sound. By requiring strategic planning and
a linking of program activities performance goals
to an organisation’s budget, decision-making and
confidence in the organisational performance is
expected to improve. A business organisation vision
is one of its most important pieces of intangible
assets. Vision is planned by strategy and executed
by values that drive day to day decision-making
(Sullivan, 2000). The economic value of an intangible
asset drives the decision to invest further, continue to
hold onto it, or dispose of it. An intangible economic
value is the measure of the utility it brings to the
business organisation. Strategy is used to develop
and sustain current and competitive advantages
for a business, and to build competitive advantages
for the future. Competitive advantage strategy
depends on the command of and access to effective
utilisation of its resources and knowledge. Strategy
is the identification of the desired future state of
the business, the specific objectives to be obtained,
and the strategic moves necessary to realise that
future. Strategy includes all major strategic areas,
such as markets, suppliers, human resources,
competitive advantages, positioning, critical success
factors, and value chains (Alter 2002). In today’s fast
changing business environment, the only way to gain
competitive advantage is by managing intellectual
capital, commonly known as knowledge management
This paper was presented at the Knowledge Management International Conference 2008 (KMICE 2008), Langkawi,
10 -12 June 2008
VOLUME Six NUMBER two july - december 2008 PLATFORM
129
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
(KM). Nowaday’s knowledge is increasingly becoming
the greatest asset of organisations (Ravi Arora, 2002).
The basic objective of a knowledge management
programme should be well understood and its
potential contribution to the business value should
be established before beginning of the process. One
of the objectives of a KM programme is to avoid reinvention of the wheel in organisations and reduce
redundancy. Secondly, a KM programme is to help
the organisation in continuously innovating new
knowledge that can then be exploited for creating
value. Thirdly, a KM programme is to continuously
increase the competence and skill level of the people
working in the organisation (Ravi Arora, 2002). KM
being a long term strategy, a Balanced Score Card
(BSC) helps the organisation to align its management
processes and focuses the entire organisation to
implement it. The BSC is a management framework
that measures the economic and operating
performance of an organisation. Without a proper
performance measuring system, most organisations
are not able to achieve the envisioned KM targets.
BSC provides a framework for managing the
implementation of KM, while also allowing dynamic
changes in the knowledge strategy in view of changes
in the organisational strategy, competitiveness and
innovation. Inappropriate performance measurement
is a barrier to organisational development since
measurement provides the link between strategies
and actions (Dixon et al., 1990). Performance
measurement as a process of assessing progress
towards achieving predetermined goals, includes
information on efficiency. In which, resources are
transformed into goods and services, the quality of
those outputs and outcomes, and the effectiveness
of organisational operations in terms of their specific
contributions to organisational objectives (Dilanthi
Amaratunga, 2001). This paper identifies the balanced
scorecard developed by (Kaplan and Norton, 1992,
1996a) as a leader in performance measurement and
performance management in an attempt to identify
an assessment methodology for organisational
processes.
130
BALANCED SCORE CARD
Robert S. Kaplan and David P. Norton (1992) devised
the Balanced Scorecard of its present form. They
framed the balanced scorecard as a set of measures
that allowed for a holistic, integrated view of the
business process so as to measure the organisation’s
performance. The scorecard was originally created
to supplement “traditional financial measures with
criteria that measured performance from three
additional perspectives – those of customers, internal
business processes, and learning and growth”. The
BSC retained traditional financial measures. But
these financial measures tell the story of past events,
an adequate story for those companies for which
investments in long-term capabilities and customer
relationships are not critical for success. These financial
measures are inadequate, however, for guiding and
evaluating the performance of the modern companies
as they are forced by intense competition provided
in the environment, to create future value through
investment in customers, suppliers, employees,
processes, technology, and innovation. Non-financial
measures, such as customer retention, employee
turnover, and number of new products developed,
belonged to the scorecard only to the extent that they
reflected activities an organisation performed in order
to execute its strategy. Thus, these measures served
as predictors of future financial performance.
In due course of time, the whole concept of
the balanced scorecard evolved into a strategic
management system forming a bridge between
the long-term and short-term strategies of an
organisation. Many companies readily adopted the
BSC because it provided a single document in which
the linkages of activities – more particularly by giving
adequate importance to both tangible and nontangible factors – were more vividly brought out than
in any other process adopted. Clearly, opportunities
for creating value shifted from managing tangible
assets to managing knowledge-based strategies
that deployed an organisation’s intangible assets:
customer relationships, innovative products and
services, high quality and responsive operative
processes, information technology and databases
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
and employee capabilities, skills and motivation.
The BSC has grown out of itself from being just a
strategic initiative to its present form of a Performance
Management System. The balanced scorecard, as it is
today, is a Performance Management System that can
be used by organisations of any size to align the vision
and mission with all the functional requirements of
day-to-day work. It can also enable them to manage
and evaluate business strategy, monitor operational
efficiency, provide improvements, build organisation
capacity, and communicate progress to all employees.
Hence, it is being adopted by many companies across
the world today cutting across the nature of the
industry, types of business, geographical and other
barriers.
Kaplan & Norton (1992) described the Balanced
Scorecard as a process which “moves beyond a
performance measurement system to become the
organising framework for a strategic management
system”. It is important that the scorecard be seen not
only as a record of results achieved, and it is equally
important that it be used to indicate the expected
results. The scorecard in this way will serve as a way to
communicate the business plan and thus the mission
of the organisation. It further helps to focus on critical
issues relating to the balance between the short and
long run, and on the appropriate strategic direction
for everyone’s efforts (Olve et al., 1999). The BSC
allows managers to look at the business from the four
perspectives and provides the answers to the above
basic questions, as illustrated in Figure 1.
Customer Perspective
This perspective captures the ability of the organisation
to provide quality goods and services, the effectiveness
of their delivery, and overall customer service and
satisfaction. Many organisations today have a mission
focused on the customer and how an organisation
is performing from its customer’s perspective has
become a priority for top management. The BSC
demands that managers translate their general
mission statement on customer service into specific
measures that reflect the factors that really matters
to their customer. The set of metrics chosen under
this perspective for this study were: enhance market
share by 5%, 10% increase in export sales, obtain
competitive pricing, and increase after sales service
outlets by 10%.
Internal Business Processes Perspective
The business processes perspective is primarily an
analysis of the organisation’s internal processes. Internal
business processes are the mechanisms through
which performance expectations are achieved. This
perspective focuses on the internal business process
results that lead to financial success and satisfied
customers expectations. Therefore, managers need
to focus on those critical internal business operations
that enable them to satisfy customer needs. The set
of metrics chosen under this perspective for this study
were: improving productivity standards, eliminating
Financial Perspective
“How do we look to our
shareholders?”
Customer
Perspective
Strategy
“How do our
customers see us?”
Internal business
processes
Perspective
“What must we
excel at?
Learning and Growth
Perspective
“How can we continue to
improve?”
Source. Kaplan and Norton (1996 a)
Figure 1. Balanced Score Card
VOLUME Six NUMBER two july - december 2008 PLATFORM
131
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
defects in manufacturing, provide adequate technical
knowledge and skill for all the levels of employees and
customer feedback to be integrated in the operation.
Learning and Growth Perspective
The targets for success which keep changing in the
intense competition requires that organisations make
continual improvements to their existing products
and processes and have the ability to introduce
entirely new processes with expansion capabilities.
This perspective looks at such issues, which includes
the ability of employees, the quality of information
systems, and the effects of organisational alignment in
supporting accomplishment of organisational goals.
The set of metrics chosen in this study under this
perspective were: involve the employees in corporate
governance, become a customer driven culture and
inculcate leadership capabilities at all levels.
According to the Balanced Scorecard Collaborative,
there are four barriers to strategic implementation:
1. Vision Barrier – No one in the organisation
understands the strategies of the organisation.
2. People Barrier – Most people have objectives that
are not linked to the strategy of the organisation.
3. Resource Barrier – Time, energy, and money are
not allocated to those things that are critical to the
organisation. For example, budgets are not linked
to strategy, resulting in wasted resources.
4. Management Barrier – Management spends too
little time on strategy and too much time on
short-term tactical decision-making.
All these observations call for not only developing
proficiency in formulating an appropriate strategy
to make the organisational goals relevant to the
changing environment but also call for an effective
implementation of the strategy.
Financial Perspective
Financial performance measures indicate whether
the organisation’s strategy, implementation, and
execution are contributing to bottom line
improvement. It shows the results of the strategic
choices made in the other perspectives. By making
fundamental positive changes in their operations, the
financial numbers will take care of themselves. The set
of metrics chosen in this study under this perspective
were: 12% return on equity to be achieved, 20%
revenue growth, 2% reduction in cost of capital and
7% reduction in production cost.
METHODOLOGY
This paper used the Balanced Score Card approach
proposed by (Robert Kaplan and David Norton, 1992)
and the model adopted by (Brown and Gibson, 1972)
along with the extension to the model provided by
(Raghavan and Punniyamoorthy, 2003) to arrive at a
single measure called Effectiveness Score (ES). This
score was used to compare the differences between
targeted performance and actual performance of any
organisation.
132
EFFECTIVENESS SCORE FOR THE BALANCED
SCORECARD (ESBSC)
The Balanced Scorecard in its present form certainly
eliminates uncertainty to a great extent as compared
to the traditional financial factors based performance
measurement systems. However when this study set
out to measure the actual performance against the
targeted performance, most of the criterions was met.
For some factors, actual performance was greater
than the targeted performance, and for others, it was
less. Therefore for the decision makers there might
be some kind of confusion regarding the direction in
which the organisation is going. That is, the decision
maker might not be clear whether the firm is improving
or deteriorating. This is because the firm might have
achieved the desired performance in the not-so-vital
parameters but would have failed to show required
performance in many vital parameters. Hence, it
becomes imperative to provide weightage for the
factors considered, so as to define the importance to
be given to the various parameters. This will provide
a clear direction to the management to prioritise the
fulfillment of targets set for those measures which
were ascribed a larger weightage.
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
The organisation can reasonably feel satisfied if it
were able to achieve the targets set for it, as it would
encompass all the performance measures. Basically,
“The Balanced scorecard” wass constructed by taking
into account all the strategic issues. The effectiveness
score, which this study suggests, is basically derived
for the balanced score card. The single benchmark
measure “The Effectiveness Score for the Balanced
Scorecard” created would clearly mean that the firm
will be reasonably be in a position to evaluate the
achievement of the strategic targets. In short, it is a
single benchmarking measure, which evaluates underor over-achievement of the firm in respect of fulfilling
the goals set by the organisation. It can also provide
variations of the actual measure from the targeted
measure under each of the factors considered. Thus
the framework suggested in this paper will provide a
single benchmark information for the decision makers
to take appropriate action and concentrate on such
measures which would result in the achievement of
the strategic needs of the company.
four perspectives. The perspectives, the measures
under each perspective, the target and actual values
of each measure were analysed in a framework as
shown in Figure 2.
The Target Performance (TP) and Actual Performance
(AP) were calculated using the following method:
Balanced score for Balanced scorecard
(Target Performance)
= a1(b1c1+b2c3+b3c5) + a2(b4c7+b5c9)
+a3(b 6 c11+b7c13) + a4(b8c15) (1)
Balanced score for Balanced scorecard
(Actual Performance)
= a1(b1c2+b2c4+b3c 6) + a2(b4c8+b5c10)
+a3(b 6 c12+b7c14) + a4(b8c16) (2)
There are four Levels in the Effectiveness Score for
Balanced Scorecard Model.
Level I: The first level is the goal of the model.
DEVELOPMENT OF EFFECTIVENESS SCORE (ES)
Let us now see the development of Effectiveness Score
Model for Balanced Scorecard. As discussed earlier
the Balanced Scorecard divides all the activities under
Level II: This level consists of the criteria for evaluating
organisational performance under the following
categories:
Effectiveness Score for Balanced Scorecard
I-Level
Goal
II – Level
Criteria
Financial
Perspective
(a1)
Customer
Perspective
(a2)
Internal Business
Perspective
(a3)
Learning& growth
Perspective
(a4)
III- Level
Sub Criteria
b1
b2
b3
b4
b5
b6
b7
b8
IV- Level
Alternatives
c1 c2 c3 c4 c5 c6
TP AP TP AP TP AP
c7 c8 c9 c10
TP
AP
TP
AP
c11 c12 c13 c14
TP
AP
TP
AP
c15 c16
TP
AP
Figure 2. Framework for calculating the Balanced Score for Balanced Score Card
VOLUME Six NUMBER two july - december 2008 PLATFORM
133
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
•
•
•
•
Financial Perspective (a1)
Customer Perspective (a2)
Internal Business Process Perspective (a3)
Learning and Growth Perspective (a4)
Level III: Each Perspective may have sub-criteria for
measuring organisational performance. To measure
each criterion or sub-criteria measures were identified.
These were referred to as bi
Level IV: For each measure, identified targets were set.
These target performance values were then compared
with the actual performance achieved. In nutshell, the
score was arrived based on the relative weightages of
the items incorporated in the model, based upon the
classification suggested in the Balanced Scorecard
approach.
The factors of level II and level III were evaluated using
the Preference theory. The relative weightage for each
factor was arrived at by pair wise comparison using the
preference theory. These factors were compared pair
wise and 0 or 1 was assigned, based on the importance
of one perspective over another. In each level the
factor’s relative weightage was established by pair wise
comparison. In the process of comparison, if the first
factor were more important than the second factor, 1
for the first and 0 for the second were assigned. If the
first factor were less important than the second factor,
0 for the first and 1 for the second were assigned. If both
perspectives were valued equally, 1 was assigned for
both perspectives. When the values were assigned, it
was seen that results of the comparison decision were
transitive. i.e., if the factor 1 were more important than
factor 2 and factor 2 were more important that factor
3, then the factor 1 was more important than factor 3.
The factors of level IV were grouped into financial and
non-financial factors to measure the effectiveness of
the organisation’s activity. The financial factors were
cost and benefit. Non-financial factors were classified
into factors related with time dimensions and other
factors. The above said factors were then brought
Tangible Factors
To be maximized
Financial Factors
Monetary
Dimensions
-Labour saving
-Material saving
-Inventory Cost
saving
To be minimized
Non Financial
Factors
Time
Dimensions
Utilization
Time
Financial Factors
Others
Productivity
Figure 3. Framework for calculating Level IV – Alternatives
134
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Monetary
Dimensions
-Labour Cost
-Material
Cost
-Overheads
Non Financial
Factors
Time
Dimensions
-Cycle time
-Set up time
Others
Loss
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
under categories, those which were maximised, and
the factors which were minimised.
A general expression was then framed, considering
the entire set of factors. The expression was framed
in such a manner that the factors were converted
into consistent, dimensionless indices. The sum of
each index was equal to 1. This was used to evaluate
the factors in order to assist to arrive at the relative
weightage at the lowest level. This was the framework
developed by (Ragavan, 2003).
ESI = BM I (1/ Σ BM) + [CM I Σ (1/CM)]-1
+ BT I (1/Σ BT) + [TM I Σ (1/TM)]-1
(3)
+ NFI (1/Σ NF) + [NFM I /Σ 1/NFM]-1 Where
ESI = Effectiveness score for alternative ‘I’
BM I = Benefit in money for alternative ‘I’
BT I = Benefit in time for alternative ‘I’
CM I = Cost to be minimised for alternative ‘I’
TM I = Time to be minimised for alternative ‘I’
NFI = Non financial factors for alternative ‘I’ to be
maximised
NFM I =Non financial factors for alternative ‘I’ to be
minimised
The relative weightage for all the factors were arrived
and the Effectiveness Score for the Balanced Scorecard
was arrived using equations (1) and (2) for sample
framework given in Figures 2 and 3. How the company
had fared was made by comparing the figures of
targeted performance and the actual performance.
LIMITATION OF THE RESEARCH
Level II and Level III factors were evaluated using
preference theory which has certain limitations. In the
comparison of the degree of importance of one factor
over another and by assigning 1 for a factor and 0 for
another, meant that 0 importance is attached to that
factor. There is a possibility, that the factor be uniformly
0 in value in all the pair wise comparisons. This would
result in a factor getting 0 relative importance. In
other words, in a decision, a 0 value factor does get a
role, which is not necessary.
FUTURE PROSPECT OF THE RESEARCH
To remove the abovesaid limitation, future research
may be carried out to evaluate criteria (ai ’s) and the
sub-criteria (bi ’s) by using Analytic Hierarchy Process
(AHP). Even by adopting AHP, pair wise comparisons
can be made and different values be assigned based
on the degree of importance ranging from 1-9. The
reciprocal values can also be assigned based on
the importance of one factor over the other. This
may provide further refinement towards adequate
weightages to the relevant criteria and sub-criteria.
CONCLUSION
Knowledge management, as a long term strategy,
could use BSC to help the company align its
management processes which focuses the entire
organisation in its implementation. Without a proper
performance measuring system, most organisations
were not able to achieve the envisioned KM targets.
BSC provided a framework for managing the
implementation of KM while also allowing dynamic
changes in the knowledge strategy in view of changes
in the organisational strategy, competitiveness and
innovation. There were many attempts made to show
the efficacy of the usage of the balanced scorecard
for showing better performance. While retaining all
the advantages that were made available by using the
balanced score card approach in providing a frame
work for showing better performance, a process of
calculating a benchmark figure called “Effectiveness
score” was added for the analysis. This study identified
parameters whose actual performance varied from
the targeted performance and found their relative
proportion of adverse or favourable contributions
to the performance of the company by assigning
appropriate weights for such parameters whether
financial or non-financial. Therefore, this study was
in the position to objectively capture the reason for
variations in the performance from the targeted
levels in all the functional areas of the business with
the use of the concepts of balanced scorecard as well
as applying the extended information arising out of
arriving at the “Effectiveness score for the balanced
scorecard”. In conclusion, arriving at this score by
VOLUME Six NUMBER two july - december 2008 PLATFORM
135
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
and large is considered a powerful approach in
formulating a business excellence model. This would
certainly help users of this approach to make an
objective evaluation while implementing the same in
their business environment.
REFERENCES
[1]
Alter, S. (2002). Information system the foundation of
E-business (4th ed.): Prentice-Hall, Upper Saddle River, NJ.
[2]
Brown, O. P. A., & Gibson, D. F. (1972). A quantified model
for facility site selection application to the multi product
location problem, AIIE Transaction, 4, 1, 1-10.
[3]
Dilanthi A., David B., & Marjan S. (2001). Process
improvement through performance measurement: the
balanced scorecard methodology, Work Study, 50, 5, 179188.
[4]
Dixon, J. R., Nanni, A. J., & Vollman, T. E. (1990). The new
performance challenge: Measuring operations for world
class competition: Business One Irwin, Homewood, IL.
[5]
Kaplan, R. S., & Norton, D. P., (1992). The balanced score card,
Measures that drive performance, Harvard Business Review,
January, 71-79.
[6]
Kaplan, R. S., & Norton, D. P., (1992). The balanced score card:
Translating strategy into action, Harvard Business School
Press, Boston, M.A.
[7]
Kaplan, R. S., & Norton, D. P., (1996a). The balance score card,
Harvard Business School Press, Boston, MA.
[8]
Olve, N., Roy, J., & Wetter, M. (1999). Performance drivers: A
practical guide to using the balanced scorecard: John Wiley
& Sons, Chichester.
[9]
Ragavan, P. V., & Punniyamoorthy, M. (2003). Strategic
decision model for the justification of technology selection,
International Journal of Advanced Manufacturing
Technology, 21, 72-78.
P. D. D. Dominic obtained his MSc degree
in operations research in 1985, MBA from
Regional
Engineering
College,
Tiruchirappalli, India in 1991, Postgraduate
Diploma in Operations Research in 2000
and completed his PhD in 2004 in the area
of job shop scheduling at Alagappa
University, Kariakudi, India. Since 1992 he
has held the post of Lecturer in the
Department of Management Studies, National Institute of
Technology (formerly Regional Engineering College),
Tiruchirappalli- 620 015, India. Presently, he is working as a Senior
Lecturer, in the Department of Computer and Information
Sciences, University Teknologi PETRONAS, Malaysia. His fields of
interest are Information Systems, Operations Research, Scheduling,
and Decisions Support Systems. He has published technical
papers in international and national journals and conferences.
[10] Ravi, A. (2002). Implementing KM – a balanced score card
approach, Journal of knowledge management, 6, 3, 240249.
[11] Sullivan, P. (2000). Value driven intellectual capital: How to
convert intangible corporate assets into market value, John
Wiley & Sons, New York, NY.
136
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
A Conceptual Framework for Teaching Technical
Writing Using 3D Virtual Reality Technology
Shahrina Md Nordin*, Suziah Sulaiman, Dayang Rohaya Awang Rambli,
Wan Fatimah Wan Ahmad, Ahmad Kamil Mahmood
Universiti Teknologi PETRONAS, 31750 Tronoh, Perak Darul Ridzuan, Malaysia.
*[email protected]
Abstract
This paper presents a conceptual framework for teaching technical writing using 3D virtual reality technology
to simulate a contextual learning environment for technical learners. The 3D virtual of a virtual environment of
the offshore platform is proposed to provide learners with an opportunity to acquire effective communication
skills in their target workplace community. The goal of this project is to propose a virtual environment with realworld dynamic contents to develop effective technical writing. The theories and approaches to teaching and
learning underlying the conceptual framework of the virtual reality (VR) will be discussed. The architecture and
the technical aspects of the offshore platform virtual environment will also be presented. These include choices
of rendering techniques to integrate visual, auditory and haptics in the environment. The paper concludes with
a discussion on pedagogical implications.
Keyword: Virtual reality, offshore platform, technology, 3D
Introduction
One of the most challenging tasks for language
instructors is to provide practical learning
environments for the students. In recent years, Virtual
reality (VR) and 3D virtual learning environment (3D
VLE) have become increasingly explored. VR, through
simulations, could help overcome the limitations
of language learning within the four walls of the
classroom. The potential of VR in education however
is exploited only quite recently by educators and
institutions [1]. [2] defines VR as “an experience in
which a person is surrounded by a three-dimensional
computer-generated representation, and is able
to move around in the virtual world and see it from
different angles, to reach into it, grab it and reshape
it” [3].
3D VLE is a learning and teaching program that makes
use of a multi-user virtual environment or a single
user virtual environment to immerse students in
educational tasks [4]. The interactive and 3D immersive
features of a built virtual environment would provide
the learners a rich, interactive and contextual setting
to support experiential and active learning. This paper
therefore aims to present the conceptual framework
of a 3D VR to simulate environments of the offshore
oil platform for technical learners. Such simulation
and virtual environment of the offshore platform
will provide learners with an opportunity to acquire
relevant effective technical communication skills in
their target workplace community. The paper will
thus first discuss the underlying theories and recent
research work on virtual environment in education.
The conceptual framework for implementation of
This paper was presented at the Information Technology Symposium 2008 (ITSim08), Kuala Lumpur
25 - 29 August 2008
VOLUME Six NUMBER two july - december 2008 PLATFORM
137
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
VR to second language technical writing learners will
also be presented. The paper will also present the
basic architecture and the technical aspects of the
offshore platform virtual environment. The discussion
includes choices of rendering techniques to integrate
visual, auditory and haptics in the environment. The
paper concludes with a discussion on pedagogical
implications.
Literature Review
Many attempts have been made to develop immersive
learning environments [5, 6, 7, 8]. [6] highlighted
several memorable yet still ‘living’ environments that
include applications for medical students studying
neuroanatomy developed as early as in the 1970s,
and a system for beginners learning a particular
language [9]. The goal for developing such virtual
worlds is to facilitate learning since the environment
enables self-paced exploration and discovery to take
place. A learner can engage in a lesson that is based
on learning by doing and is able to understand the
subject matter in context. Among the advantages of
using 3D VLE for teaching and learning include the
sense of empowerment, control and interactivity; the
game-like experience, heightened levels of motivation;
support visual learners; allow self-awareness, support
interaction and enable real-time collaboration and
ability to situate students in environments and context
unavailable within the classroom [4].
From the educational perspective, virtual
environments are considered as a visualisation tool
for learning abstract concepts [6]. Touch the Sky,
Touch the Universe [5] is a virtual environment that
represents a dynamic 3-D model of a solar system.
It aims towards providing students with a more
intuitive understanding of astronomy and contributes
to the development of essential visual literacy and
information-processing skills. Another example that
emphasises virtual environment as a visualisation
tool is NICE [7], an immersive participatory learning
environment for young learners aged 6 to 10. The
application enables the children to grow plants by
manipulating variables such as water and light in the
virtual world. This environment supports immediate
138
visual feedback to the learners and provides learning
experience visualising complex models of ecological
systems. Among other examples of similar focus
are the virtual worlds in ScienceSpace [8], and Water
on Tap [10]. The ScienceSpace project consists of
three virtual worlds: NewtonWorld, MaxwellWorld,
and Pauling World; all of which allow students to
investigate concept of Physics. On the other hand,
Water on Tap is a Chemistry world where students
learn about the concepts of molecules, electron, and
atoms. In general, most virtual worlds developed rely
mainly on the visual representation in order to learn
complex and abstract scientific concepts.
In developing a virtual world with real-world dynamic
contents, visual cues should be complemented with
other modalities such as sound and touch to create an
immersive environment. [8] applied both elements of
sound and touch to complement visual in the virtual
worlds. They argued that multisensory cues assisted
students’ experience phenomena and focus their
attention to important factors such as mass, velocity,
and energy. In all three virtual worlds developed,
students interact with the environments using a 3-Ball,
a three-dimensional haptic mouse. The vibration from
the mouse is an indication for a tactile cue whereas an
audible chime is provided for the auditory. During the
interactions, tactile and visual cues were utilised to
emphasise the potential energy concept, and auditory
and visual cues to make velocity more salient. Another
example of virtual world that supports multimodal
cues for interactions is The Zengo Sayu [9], a virtual
environment for Japanese Language Instruction.
This system allows the use of combined voice and
gesture recognition in an educational setting. It was
designed to teach Japanese prepositions to students
who have no prior knowledge of the Japanese
language. Students can hear digitised speech
samples representing the Japanese name of many
virtual objects and their relative spatial location when
touched by the user in the virtual environment. Even
though multimodal cues assist in creating a sense of
presence in a virtual environment, such integration
needs careful consideration as some modalities such as
touch sensation is very context dependent [10]; thus,
a reason for touch modality being scarcely considered
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
for a virtual environment. However, in order to obtain
a dynamic virtual environment, there is a need to take
into account other modalities besides visual so that a
more meaningful interaction similar to that in the real
world could be supported.
Virtual reality environments have in fact generated
high interest amongst language professionals. It
is argued that immersion of the target language is
essential to learn a language and simulations through
virtual reality offers such learning experience. Thus,
a number of VR environments for language learning
have been constructed. As mentioned earlier, an
example of language learning environment is Zengo
Sayu [9], developed by Howard Rose. The VR Japanese
language learning environment allows students to
interact with objects which talk.
In the context of language learning classroom, the
most beneficial characteristic about 3D virtual
environments is that they provide a first-person form
of experiential learning [18]. Most lessons today
however are based on textbooks, which the knowledge
is based on a third-person’s knowledge. It was further
argued [18] that the qualitative outcomes of thirdperson versus first-person learning are different in that
in the case of the former, the learning outcomes are
shallow and retention rates are usually low. He further
argued that through virtual reality, learners learn
language through own experiences with autonomy
over their own learning.
In the quest for authentic learning materials to be
used in a language classroom, [11] examined three
virtual zoos on the net for language learners to
explore. It is argued that by using authentic learning
materials, it may make a personal connection with the
learners. Language learning is thus presented in a
contextualised manner where the subject matter can
be meaningful and relevant to the learners. [12] also
advocated the use of authentic texts to teach language
skills provides learners with real-life opportunities to
use the target language.
In the case of technical writing learners, the use
of authentic language, in the context of technical
communication which is made highly relevant and
personal to the learners, would seem to be best
represented in the form of a simulation of their future
workplace environment that is the oil platform.
With this in mind, the researchers seek to design a
conceptual framework for teaching technical writing
using 3D virtual reality technology at the offshore oil
platform.
The Conceptual Framework
A line of research indicates that language learning
should be facilitated by pedagogically sound
instruction [11]. It is further argued that learning
activities are supposed to be contextualised and
to some extent relevant to the learners [12]. A
technical writing course that aims to help learners
write the different kinds of technical writing is thus
contextualised and tailored to prepare them for
their future professional discourse community. The
underlying rationale in a technical writing course is
that the learners need to learn specific ways to write
(e.g. memorandums, proposal, reports and technical
descriptions) to participate effectively in their future
working environment as engineers. Writing therefore,
in this course, stresses on the notion that writing varies
with the social context where it is produced, that we
have many kinds of writing with different situations
[13]. With this in mind, the learners were exposed
to good models of many types of technical writing
as materials. As pointed out by [14], “learners are
often found incapable to replicate the expert generic
models due to their communicative and linguistics
deficiencies” (cited in [13]:370). It was found motivating
to the learners to be exposed to “good ‘apprentice’
generic exemplars, which can provide a realistic model
of writing performance for undergraduate students”
[13].
Such an approach to teaching technical writing,
though contextualised within the learners’ domain,
however, has its limitations. Language practitioners do
not only face the challenge to make learning relevant
to the students’ needs in their future workplace
but also to provide learners with practical learning
environments. There has indeed been a limited
VOLUME Six NUMBER two july - december 2008 PLATFORM
139
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
Exposure to Different
type of writing
opportunity to acquire effective communication skills
in their target workplace community. Figure 1 shows
the proposed framework.
Classroom activity
Lecture
Student-Centered Activities
Exposure to Relevant Virtual
Environment
Language lab activity
• Collaborative virtual
walkthrough
• Performed Assigned
Tasks
• Selection of report type
Report Writing & production
Classroom activity
• Planning
• Drafting
• publishing
Report Evaluation
Classroom activity
Figure 1
number of researches reported in exploiting human
capabilities mainly the visual, sound and touch senses
to create such environments. VR through simulations
could help overcome the limitations of a technical
writing course.
As many of the students in this institution will serve
at the various local and international oil and gas
companies, the proposed conceptual framework
utilises a simulated environment of the offshore
oil platform to provide the contextual learning
environment for teaching technical writing to these
learners. The virtual environment of the offshore
platform is expected to provide the learners an
140
In the initial phase of the framework, learners would
first be exposed to the type of writing required in their
future workplace as a classroom activity. Following the
familiarisation of different types of technical writings,
these engineering students would then be presented
with the relevant 3D virtual world, in this case an
oil platform. This second phase would take place
in the English language laboratory equipped with
computers whereby each student would be assigned
to a computer. Despite this individual assignment,
students would be working collaboratively, sharing
the same virtual world and interacting with each
other. As these students walk around the virtual oil
platform they could talk to other students via their
avatars’ representations, who would assume different
roles e.g. the supervisor, mechanical engineer, civil
engineer etc. They may interact and discuss with one
another on the task assigned to them by the instructor.
For example, there maybe equipment malfunction
at the platform and the student would be required
to identify the problem and write an equipment
evaluation report to the supervisor. The learners thus
need to communicate and discuss with one another,
whoever assumed the different roles.
Learners need to relate the purpose of writing to the
subject matter, the writer/audience relationship and
the mode or organisation of the text. This requires
the learners to examine their surrounding and the
current social context. This thus would offer learners a
view of how different texts are written in accordance
to their purpose, audience and message [15]. The VR
application would provide students with note-taking
functionalities to enable them to collect and write
relevant information and comments during their
virtual walkthrough. These notes should be easily
accessible later when the students prepare their report.
Students were expected to decide the type of report
to write during exposure to the virtual environment.
After being exposed to the organisation and language
used in the texts, learners then go through a multipledrafts process, again as a classroom activity. Instead of
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
turning in a finished product right away, learners will
be asked for multiple drafts of a work. Rewriting and
revision are integral to writing, and editing is an ongoing multi-level process, which consists of: planning,
drafting and finally publishing the end product – the
report. The report would then be submitted to the
supervisor (a role assumed by a student) who would
read and examine the report. Figure 2 illustrates the
main features in the teaching of writing equipment
evaluation report.
Situation
(The equipment to be examined
through 3-D VR software)
Examining Purpose
(To evaluate the equipment e.g.
temperature and make relevant
recommendation)
Consideration of mode/field/tenor
(Internal report, data for technical evaluation,
audience)
In-class activities:
Planning
Drafting
Publishing
Equipment
Evaluation
Figure 2
User
User
User
Desktop computer
Desktop computer
Desktop computer
OPVE
Graphics,audio,
text, gestures
Graphics,audio
, text, gestures
Computer server
User
User
User
Desktop computer
Desktop computer
Desktop computer
Figure 3
Offshore Platform Virtual Environment
Architecture
The aim of developing an offshore platform virtual
environment (OPVE) is to provide a simulation of
the actual workplace where learners could acquire
effective communication skills through experiential
and active learning. As such, it is necessary to develop
a virtual world that closely mimics the real offshore
platform environment. Considering the complexity
of modeling this environment using 3D modeling
software, the following approach is proposed. A
series of photographs of the oil platform environment
will be taken using a digital camera and a Novoflex
VR-system PRO Panaroma Adaptor. The later allows
multi-row and spherical stitched pictures and
panaroma views to be created using a software called
Easypano Tourweaver Professional. This software
allows users a virtual walkthrough and exploration
of the 3D realistic images of the offshore platform
environment. Hotspots will be created for ease of
navigation and for users to quickly move to specific
locations. Information and instruction texts as popups are provided to further assist student explore
and perform the assigned tasks. The OPVE resides in
a computer server and can be accessed by students
via their computer desktops. These concepts could
be extended to multi-user collaborative work to allow
several users to share the OPVE, create representations
of the users, communicate and interact with each
other via graphics, audio, text and gestures in realtime over a network (Figure 3).
VOLUME Six NUMBER two july - december 2008 PLATFORM
141
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
Pedagogical Implications
There are several pedagogical implications that can be
drawn from this paper. First of all, it is acknowledged
that much of current research work in the design of
VR focuses on the technology itself and laboratory
prototypes. This paper however proposes a conceptual
framework for the use of VR as a tool in the teaching
and learning of technical writing. The increasing use
of Information and Communication Technologies
(ICT) in the education field over the last decade has
put pressure on second language (L2) writing teachers
to embrace what computers and technology offer.
Technology has its place and has quite an impact on
L2 writing instruction. As reminded by [16], language
and communication “…teachers should be prepared
to bring computers into the centre of their own
pedagogical practice” [16]. New technology-based
pedagogies have been increasingly integrated into L2
writing instruction that is believed could improve the
learners’ communication skills.
The obvious difference between the 3D VR simulation
and the traditional textbook is that the 3D environment
provides students with learning activities that are very
much action-oriented. It allows the learners to directly
experience for themselves the thing they seek to learn
[17]. Such approach provides the learners with a ‘firstperson form of learning’ instead of using solely text
books in traditional classroom [18]. Such approach
is very much in line with the task-based approach in
teaching which is propagated by [19]. In such taskbased teaching, students are required to engage in
interaction in order to fulfill a task. [20] stated that
“The role of tasks has received further support from
researchers in second language acquisition who were
interested in developing pedagogical application of
second language acquisition theory” (cited by [21]:
223). Task-based learning, as proposed in this paper,
focuses on the process of learning and problem-solving
as it is organised around a set of tasks that are related
to real life. [22] defines task as a “piece of meaningfocused work involving learners in comprehending,
producing and/or interacting in the target language,
and that tasks are analysed or categorised according
to their goals, input data, activities, settings and
142
roles”. The tasks allow interactions between learners
and the linguistic environment, as the learners focus
on participating in the tasks actively to solve the
“real-world problem” or the “pedagogic problems”.
For example, if there is an equipment malfunction,
the learners would need to identify the problem
and come up with an equipment evaluation report.
In this way, learners could put writing activities in a
relevant context which can facilitate and enhance
internalisation of target language and report writing
skills.
[20] further argued that “the underlying language
systems” will develop while the students carry out the
task. The language systems acquired are within the
specific genre in the context of real-world oil platform.
The acquisition of such relevant language system and
genre is crucial to enable students to produce a report
in such contextual setting. According to [23], genre
referred to “abstract, socially recognised ways of
using language”, which is used by the students’ future
professional community (engineers and technical
workers at the oil platform). As argued by [24] “a
virtual environment to learn the target language
would be very helpful for the learner to acquire the
needed knowledge for communication in the social
and cultural sense” – in the case of this paper, the
social and the cultural environment of technical
engineers at the oil platform. This would also help to
minimise the gap between classroom activities and
the professional setting.
Another pedagogical implication is in the use of the
teaching materials. For a long time, education has
relied on pictures and drawings to describe a variety
of objects or mechanisms. The graphics used in VR
can be of great assistance in the teaching and learning
activities. [25], for example, proved the effectiveness
of using software visualisation and animation in
teaching robotics and describing mechanical systems
to undergraduate engineering students.
Such
simulation-based learning overcomes the limitations
in using textbooks. In contrast to pictures in textbooks,
students in classrooms using VR are presented with
virtual objects in natural situations supporting the
communicative tasks they need to perform. When
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
learners are very much involved in the materials that
they learn this leads to a more effective learning to
take place as the traces of information would be more
vivid in the minds.
Conclusion
Teaching using 3D graphics has evolved as there
is a tremendous advancement in the hardware
and software technology. VR offers potentials in
education. Teaching and learning activities can be
greatly facilitated by the use of VR which helps to
minimise the gap between the classroom and the
professional world. Since the paper presented here
offers only the conceptual framework, researchers
and educators should perhaps conduct an empirical
study on its implementation in a technical writing
classroom. Such empirical study could evaluate the
effectiveness of such approach to teaching technical
communication by testing the students’ performance
in tests. Students’ performance in a traditional
classroom situation may be compared to the students’
performance in the classroom using VR. Future
research could also look into student’s perception of
such approach through a more qualitative study.
References
[1]
Manseur, R. “Virtual Reality in Science and Engineering
Education”, In Proceeding of the 35th ASEE/IEEE Frontiers in
Education Conference, Indianapolis, IN. (2005)
[2]
Rheingold, H. (1991). Virtual Reality. New York, NY: Summit
[3]
Jung, H. J. “Virtual Reality for ESL Students”, The Internet
TESL Journal, VIII, (10), (2002)
[4]
Nonnis, D. 3D Virtual Learning Environments, Educational
Technology Division, Ministry of Education, Singapore
2005, pp. 1-6
[5]
Yair, Y., Mintz, R. & Litvak, S. “3D-Virtual Reality in Science
Education: An Implication for Astronomy Teaching”, Journal
of Computers in Mathematics and Science Teaching, 20 (3),
(2001), pp. 293-305
[6]
Dean, K. L., Asay-Davis, X. S., Finn, E. M., Foley, T., Friesner, J.
A., Imai, Y., Naylor, B. J., Wustner, S. R., Fisher, S. S. & Wilson,
K. R. “Virtual Explorer: Interactive Virtual Environment for
Education”, Presence, (2000), 9(6), pp. 505-523
[7]
Roussos, M., Johnson, A., Moher, T. Leigh, J. Vasilakis C. &
Barnes, C. “Learning and Building Together in an Immersive
Virtual World”, Presence, 8 (3), (1999), pp. 247-263
[8]
Dede, C., Salzman, M. C. & Loftin, R. B. “ScienceSpace: Virtual
Realities for Learning Complex and Abstract Scientific
Concepts”, In Proceedings of IEEE Virtual Reality Annual
International Symposium, New York: IEEE Press, (1996), pp.
246-253
[9]
Rose, H. & Billinghurst, M. “Zengo Sayu: An Immersive
Educational Environment for Learning Japanese” (Technical
Report), Seattle: University of Washington, Human Interface
Laboratory of the Washington Technology Center, (1996)
[10] Byrne, C. M. (1996), “Water on Tap: The Use of Virtual Reality
as an Educational Tool”, Unpublished doctoral dissertation,
University of Washington, Seattle,WA
[11] LeLoup, J. W. & Ponterio, R. “On the Net”. Language Learning
and Technology , 9 (1), (2005), pp. 4-16
[12] Hadley, O. A. Teaching Language in Context. Boston, MA:
Heinle & Heinle, (2001).
[13] L. Flowerdew, “Using a genre-based framework to teach
organizational structure in academic writing”, ELT Journal,
54 (4), (2000), pp. 369-378
[14] Marshall, S. “A genre-based approach to the teaching of
report writing”. English for Specific Purposes. 10, (1991). pp.
3-13
[15] Macken-Horarik, M. “Something to shoot for: A systematic
functional approach to teaching genre in secondary
school science”. In A. M. Johns (Ed.), Genre in the classroom
Mahwah, NJ: Erlbaum. (2002), pp. 21-46.\
VOLUME Six NUMBER two july - december 2008 PLATFORM
143
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
[16] Pennington, M. C. The impact of the computer in second
language writing. In Kroll, B. (Ed.) Exploring the Dynamics of
Second Language Wrting, USA: Cambridge University Press,
(2003)
[17] Winn, W. “A Conceptual Basis for Educational Applications of
Virtual Reality”. Technical Report TR-93-9, Human Interface
Technology Laboratory, University of Washington, (1993)
[18] Chee, Y. M. “Virtual Reality in Education: Rooting Learning in
Experience”. In Proceeding of the International Symposium on
Virtual Education, (2001). Busan, South Korea
Shahrina Md. Nordin obtained her PhD
from Universiti Sains Malaysia, Malaysia.
She is attached as a lecturer to Universiti
Teknologi PETRONAS, Perak, Malaysia. She
started her career as a lecturer at Institut
Teknologi Perindustrian, Johor Bahru and
was then recruited as Assistant Lecturer, at
Multimedia University, Melaka. Shahrina
has a Master’s degree in TESL from Universiti
Teknologi Malaysia, Johor and obtained her first degree, in English
Language and Literature, from International Islamic University
Malaysia.
[19] Willis, J. “Task-based Learning – What kind of adventure?”
The Language Teacher, 22 (7), (1998), pp. 17-18
Suziah Sulaiman obtained her PhD from
University College London, United
Kingdom. She is currently a lecturer at
Universiti Teknologi PETRONAS, Malaysia.
Her research interests include topics on
human computer interactions, user haptic
experience, and virtual environment.
[20] Long, M. & Crookes, G. “Three Approaches to Task-based
Syllabus Design”, TESOL Quarterly, 26 (1). (1992)
[21] Richards, J. C., and T. Rodgers. “Approaches and methods in
language Teaching”. Cambridge University Press. (2001)
[22] Nunan, D. “Designing tasks for communicative classroom.
Cambridge”. Cambridge University Press. (1989)
[23] Hyland, K. “Genre-based pedagogies: A social response to
process”. Journal of Second Language Writing 12, (2003),
pp. 17-29
[24] Paepa, D, Ma, L., Heirman, A., Dessein, B. Vervenne, D.,
Vandamme F. & C. Willemen, “A Virtual Environment for
Learning Chinese”. The Internet TESL Journal, 8 (10), (1998).
Available online http://iteslj.org/
[25] Manseur, R. “Visualization tools for robotics education”.
Proceedings of the 2004 International Conference on
Engineering Education, Gaineville, Florida, October 17-21,
(2004)
144
Dayang Rohaya Awang Rambli obtained
her PhD at Loughborough University,
United Kingdom. She is currently a lecturer
of the Computer & Information Sciences
Department of Universiti Teknologi
PETRONAS, Malaysia. vHer main interest
areas are human factors in virtual reality,
applications of VR in education & training
(interactive learning systems), and
augmented reality applications in games and education.
Wan Fatimah Wan Ahmad received her
BA and MA degrees in Mathematics from
California State University, Long Beach,
California USA in 1985 and 1987. She also
obtained Dip. Ed. from Universiti Sains
Malaysia in 1992. She completed her PhD
in Information System from Universiti
Kebangsaan Malaysia in 2004. She is
currently a senior lecturer at Computer &
Information Sciences Department of Universiti Teknologi
PETRONAS, Malaysia. She was a lecturer at Universiti Sains
Malaysia , Tronoh before joining UTP. Her main interests are in the
areas of mathematics education, educational technology, human
computer interaction and multimedia.
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
MULTI-SCALE COLOR IMAGE ENHANCEMENT
USING CONTOURLET TRANSFORM
Melkamu H. Asmare, Vijanth S. Asirvadam*, Lila Iznita
Universiti Teknologi PETRONAS, 31750 Tronoh, Perak Darul Ridzuan, Malaysia
*[email protected]
Abstract
Images captured with insufficient illumination generally have dark shadows and low contrast. This problem
seriously affects other further processing schemes like human visual perception, face detection, security
surveillance etc. In this study, a multi-scale colour image enhancement technique based on contourlet transform
was developed. Contourlet transform has better performance in representing the image salient features such
as edges, lines, curves and contours than wavelet transform because of its anisotropy and directionality. It is
therefore well-suited for multi-scale edge based colour image enhancement. The image was first converted
from RGB (red, green, blue) to a CIELUV (L is for lightness) model and then the contourlet coefficients of its L
component were adjusted to preserve the original colour using modified power law transformation function.
The simulation results showed that this approach gave encouraging results for images taken in low light and/or
non-uniform lighting conditions.
Keywords: Contrast enhancement, contourlet transform, wavelet transform, transfer function, colour space.
Introduction
Image enhancement is a digital signal processing
branch aimed at assisting image visual analysis. It is
widely used in medical, biological and multimedia
systems to improve the image quality. Theoretically,
image enhancement methods may be regarded as an
extension of image restoration methods. However, in
contrast to image restoration, image enhancement
frequently requires intentional distorting of image
signals such as exaggerating brightness and colour
contrasts, deliberate removal of certain details that
may hide important objects, converting gray scale
images into colour, etc [1]. In this sense, image
enhancement is image preparation or enrichment in
the same meaning these words have in mining. An
important peculiarity of image enhancement as image
processing is its interactive nature. The best results in
visual image analysis can be achieved if it is supported
by a feedback from the user to the image processing
system.
In outdoor scenes, we are very often confronted with
a very large dynamic range, resulting in areas which
are too dark or too light in the image. Saturation and
underexposures are common in images due to limited
dynamic range of the imaging and display equipment.
This problem becomes more common and severe
when insufficient or non-uniform lighting conditions
occurs. Current cameras and image display devices
do not have a sophisticated mechanism to sense the
large dynamic range of the scene.
There exists a number of techniques for colour
enhancement mostly working in colour spaces which
transform the Red, Green, Blue monochromatic
This paper was presented at the International Graduate Conference on Engineering and Science 2008, Johor Baharu,
23 - 24 December 2008
VOLUME Six NUMBER two july - december 2008 PLATFORM
145
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
space to the perceptually based hue, saturation and
intensity colour space. The techniques, such as simple
global saturation stretching [2], hue dependent
saturation histogram equalisation [3], and intensity
edge enhancement based on saturation have been
attempted in this space. Among others, there are
methods based on Land’s Retinex theory [4] and
non-linear filtering techniques in the XY chromaticity
diagram [5].
The multi-scale Retinex (MSR) introduces the concept
of multi-resolution for contrast enhancement. Ideally, if
an image can be decomposed into several components
in multiple resolution levels, where low pass and high
pass information are kept separately, then the image
contrast can be enhanced without disturbing any
details [6]. At the same time, image detail can also
be emphasised at a desired resolution level, without
disturbing the rest of the image information; finally,
by adding the enhanced components together, a
more impressive result can be obtained.
The Wavelet transform approach [7] consists of first
transforming the image using wavelet transform.
The wavelet coefficients at each scale are modified
using a non-linear function which is defined from the
gradient from coefficients relative to horizontal and
vertical wavelet bands. Finally, the enhanced image
is obtained by the inverse wavelet transform of the
modified coefficients.
This study is of the opinion that the wavelet
transform may not be the best choice for the contrast
enhancement of natural images. This observation
is based on the fact that wavelets are blind to the
smoothness along the edges commonly found in
images.
The contourlet framework provides an opportunity
to achieve these tasks. It provides multiple resolution
representations of an image, each of which highlights
scale-specific image features. Contourlet transform has
better performance in representing the image salient
features such as edges, lines, curves and contours
than wavelet transform because of its anisotropy
and directionality. Since features in those contourlet
146
transformed components remain localised in space
many spatial domain image enhancement techniques
can be adopted for the contourlet domain.[8].
For high dynamic range and low contrast images,
there is a large improvement by contourlet transform
enhancement since it can detect the contours and
edges quite adequately.
In this paper, a new image enhancement algorithm is
presented. The method for colour image enhancement
is based on multiscale representation of the image.
The paper has four sections. Section two discusses
the contourlet transform. Section three discusses the
enhancement technique. Section four shows some
experimental results and concludes the experiment.
Contourlet Transform
For image enhancement, one needs to improve the
visual quality of an image without distorting it. Wavelet
bases present some limitations, because they are not
well adapted to the detection of highly anisotropic
elements such as alignments in an image. Recently,
Do and Vetterli [8] proposed an efficient directional
multi-resolution image representation called the
contourlet transform. Contourlet transform has better
performance in representing the image salient features
such as edges, lines, curves and contours than wavelet
transform because of its anisotropy and directionality.
It is therefore well-suited for multi- scale edge based
colour image enhancement.
The contourlet transform consists of two steps:
the sub-band decomposition and the directional
transform [8]. A Laplacian pyramid is first used to
capture point discontinuities, followed by directional
filter banks to link point discontinuity into lineal
structure. The overall result is an image expansion
using basic elements like contour segments and is
named after it, contourlet transform. Figure 1 shows a
flow diagram of the contourlet transform. The image
is first decomposed into sub-bands by the Laplacian
transform and then each detail image is analysed by
the DFB.
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
Figure 1. Contourlet transform framework.
Figure 2. Contourlet filter bank.
Figure 2 shows the contourlet filter bank. First, multiscale decomposition by the Laplacian pyramid, and
then a directional filter bank is applied to each band
pass channel.
An example of contourlet decomposition coefficients,
four level pyramid transforms followed by 3, 4, 8 and
4 directional decomposition from fine to coarse levels
respectively is shown in Figure 3.
Image Enhancement Algorithm
For colour images, the colour is specified by the
amounts of Red (R), Green (G) and Blue (B) present.
Applying the grayscale algorithm independently to
the R, G, and B components of the image, will lead to
colour shifts in the image and is non-efficient, thus
unwanted. For enhancement purposes the RGB image
is converted to one of the perceptual colour spaces.
HSI (Hue, Saturation and Intensity), and LUV (L is for
Lightness,) are the most commonly used perceptual
colour spaces. LUV is claimed to be more perceptually
uniform than HSI [9]. The colour space conversion can
be formulated as follows.
To convert RGB to LUV, RGB was first converted to XYZ
components then XYZ to LUV.
 R  3.240479 - 1.53715 - 0.498535   X 
G  = - 0.969256 1.875992 0.041556  x Y 
  
  
 B  0.055648 - 0.204043 1.057311   Z 
u' =
4X
X + 15Y + 3Z
v' =
9Y
X + 15Y + 3Z
(1)
(2)
Figure 3. Contourlet coefficients
1

  Y 3
116   − 16 ,
Y
L =   n
 29  3  Y 
    ,
 
 3  Yn 
(
)
U = 13L u ' − u ' n
Y
 6
>  
 29 
Yn
3
Y
 6
≤  
 29 
Yn
(
(3)
3
)
V = 13L v ' − v ' n (4)
The quantities u n’, v n’ Xn , Yn , Z n were chromaticity
coordinates of a specified white object, which may
be termed as the white point. For a perfect reflecting
diffuser under D65 illuminant and 2° observer u n’
= 0.2009, v n’ = 0.4610. Xn = 0.950456, Yn = 1, Z n =
1.088754.
The most straightforward method would be to use
the existing grayscale algorithm for colour images by
applying it to the lightness component of the colour
image and leaving the chromaticity unaltered. The
contourlet transform was used to decompose the
lightness component into multi-scales for better
representation of major geometrical components of
the image like edges, contours and lines.
VOLUME Six NUMBER two july - december 2008 PLATFORM
147
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
The edges as features of an image contained useful
information that belonged to the high frequency
component. Human visual system is more sensitive
to the edges than the homogeneous regions. Thus
the sharpness of the edges should be preserved. The
contourlet transform has a better representation to
major geometrical parts, that is, edges, of images.
So it is the best candidate for edge based contrast
enhancement.
The contourlet transform is a hierarchy of resolution
information at several different scales and orientations.
At each level, the contourlet transform can be reapplied
to the low resolution sub-band to further decompose
the image. An image I can be characterised as, for
four decomposition levels:
1
AjI= A j+1I+ D
2
3
J+1I+D j+1I+D J+1I
4
+D
J+1I
(5)
The enhanced image:
ÃjI= Ãj+1I + [F1j+1(D1j+1I) + F2j+1(D2j+1I) + F3j+1(D3j+1I)+ F4j+1(D4j+1I)]
(6)
The important feature in contourlet transformed
components is that they remain localised in space.
Thus, many spatial domain image enhancement
techniques can be adopted for the contourlet
domain. The following non-linear mapping function
were applied to modify the contourlet coefficients.
Contourlet transformed images were treated by an
enhancement process to elevate the values of low
intensity pixels using a specifically designed nonlinear transfer function F, defined as:
L' =
(
L(0.75 z + 0,25)
+ (1 − L)* 0.5 * (1 − z ) +
2
)
L( 2 − z )
Figure 4. Transfer function for different values of z.
The function can be image dependent using z as
follows,
0
for L ≤ 0.2 Lmax

z =  L − 0.2 for 0.2 Lmax < L < 0.7 Lmax
1
for L ≥ 0.7 Lmax

(8)
Lmax is the maximum value of the sub-band image.
Thus each sub-band image was enhanced using
different enhancement function.
The over all system algorithm can be summarised in
Figure 5.
Original Image
R
G
B
Convert to LUV color space
L
(7)
This function is a modified power law transfer
function. It is a sum of three simple functions to
control the shape of the curve. It can be seen in
Figure 4 that this transformation largely increases the
luminance of darker pixels (regions) while brighter
pixels (regions) are less enhanced. Thus this process
serves as dynamic range compression. Therefore, the
line shape of the transfer function is important no
matter what mathematical functions are used. Simple
148
functions are applied for faster computation. The
range for z if [0 1].
U
V
Contourlet Transform L part
Modify the coefficients
Inverse contourlet transform
L1
U
V
Inverse color transform
Enhance Color Image
Figure 5. The overall algorithm in block diagram
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
Results and Discussion
The proposed algorithm has been applied to enhance
a large number of digital images for performance
evaluation and comparison with other algorithms.
Some typical results and discussions are given
below. The image was first decomposed into four
decomposition levels using pyramid transform
and 3, 2, 4, and 16 directions from fine to coarse
decompositions levels respectively.
The algorithm was applied to the image captured at
a low lighting condition in Figure 6(a). The enhanced
image has good quality with fine detail revealed and
a well balanced colour distribution.
The second test image was captured under very low
lighting condition. Figure 7(a) shows the original image
in which the background is entirely invisible. The
enhancement algorithm performed very well in these
kinds of images; it enhanced the almost completely
dark region and still preserved the illuminated area.
The third test image was a low and non-uniform
lighting condition as shown in Figure 8. It can be
seen that the brightness of the light source affects the
display of the cars and the building, where many details
are not visible. The enhancement algorithm tried to
balance the lighting condition the enhanced image
reveals an improved display of the entire image.
a)
b)
Figure 6. a) Original image b) enhanced image
a)
b)
Figure 7. a) Original image b) enhanced image
The algorithm was compared with most common
contrast enhancement methods such as, Histogram
equalisation (HE), Contrast Limited Adaptive Histogram
Equalisation (CLAHE) and the Wavelet transform (WT)
method. One objective function, called Structural
Similarity index (SSIM) [10] was used to measure the
performance of the algorithms.
Structural similarity index is a measure of structural
information change. It assumes one perfect reference
input image and measures the similarity. It uses three
comparisons: luminance, contrast and structure.
Suppose x and y are two non-negative images, the
structural similarity index is defined as follows. It
ranges from 0 to 1, where 1 is 100% similarity. The
constants C1, C2 and C3 are included for mathematical
stability.
SSIM ( x, y ) =
a)
b)
Figure 8. a) Original image b) enhanced image
2(2µ x µ y + c1 )( 2σ xy + c 2 )
2
( µ x + µ y 2 + c1 )(σ x 2 + σ y 2 + c 2 )
(9)
Figure 9 shows a sample comparison result for
enhancement methods. The image (Figure 9a) is
considered as a perfect image and Figure 9(b) is the
same image taken in low and non-uniform lighting
condition. Figures 9(c), (d), (e) and (f) are results of
VOLUME Six NUMBER two july - december 2008 PLATFORM
149
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
Table 1. SSIM comparison result.
Method
SSIM
HE
0.9861
CLAHE
0.9989
Wavelet
0.9991
Contourlet
0.9991
Original
0.1644
contourlet transform, wavelet transform, histogram
equalisation and contrast limited histogram
equalisation respectively.
Table 1 shows the SSIM measure of the images in Figure
8. The contourlet and wavelet methods gave the same
index. But for a human observer the contourlet result
is more natural and appealing.
Conclusion
The simulation results show that the approach of this
study gave encouraging results for images taken in
low light and/or non-uniform lighting conditions.
a)
b)
The algorithm is fast using simple mathematical
transformation function. Since the images were
well represented using the contourlet transform, the
process did not introduce any distortion to the original
image. The system has good tonal retention and its
performance is very good especially in low light and
non-uniform lighting conditions.
References
c)
e)
[1]
R. C. Gonzalez, R. E. Woods, “Digital Image Processing” 2nd
edition, Prentice-Hall, Inc., 2002
[2]
S. I. Sahidan, M. Y. Mashor, Aida S. W. Wahab, Z. Salleh,
H. Ja’afar, “Local and Global Contrast Stretching For Color
Contrast Enhancement on Ziehl-Neelsen Tissue Section
Slide Images“, 4th Kuala Lumpur BIOMED 2008 25–28 June
2008 Kuala Lumpur, Malaysia
[3]
Naik, S. K.; Murthy, C. A. “Hue-preserving color image
enhancement without gamut problem”, IEEE Transactions
on Image Processing, IEEE Transactions, Volume 12, Issue
12, Dec. 2003
[4]
B. Funt , F. Ciurea, and J. McCann, “Retinex in Matlab“,
Proceedings of CIC08 Eighth Color Imaging Conference,
Scottsdale, Arizona, pp. 112-121,2000
[5]
Lucchese, L.; Mitra, S. K. “A new filtering scheme for
processing the chromatic signals of color images: definition
and properties”, IEEE Workshop on Multimedia Signal
Processing, 2002
d)
f)
Figure 9. a) Perfect Image b) Image to be enhanced
c) Contourlet method d) Wavelet Method
d) Histogram Equalization f ) CLAHE.
150
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
[6]
Koen Vande Velde, “Multi-scale color image enhancement”,
in Proceedings of SPIE international Conference, image
processing, Vol. 3, 1999, pp. 584-587
[7]
F. Sattar, X. Gao, “ Image Enhancement Based on a Nonlinear
Multiscale Method using Dual-Tree Complex Wavelet
Transform”, IEEE, 2003
[8]
M. N. Do, M Veterli, “The Contourlet Transform: An efficient
Directional Multi-Resolution Image Representation”, IEEE
Transactions on Image Processing, Vol. 14, pp. 2091-2106,
2005
[9]
Adrian Ford, Alan Roberts,” Color Space Conversions”,
August 11, 1998
[10] Zhou Wang, Alan Conrad Bovik, Hamid Rahim Sheikh,
Eero P. Simoncelli, “Image Quality Assessment: From Error
Visibility to Structural Similarity”, IEEE Transactions on Image
Processing, Vol. 13, no. 4, April 2004, pp. 600-612
Melkamu H. Asmare was born in Dangla,
Ethiopia, in 1982. He received his BSc
Degree in Electrical and Computer
Engineering from Addis Ababa University,
in 2005. Currently, he is a Master’s Research
student in Electrical and Electronic
Engineering department of Universiti
Teknologi PETRONAS. His research interests
include image processing, colour science,
digital signal processing and image fusion.
Vijanth S. Asirvadam is from an old mining
city of Malaysia called Ipoh. He studied at
University Putra, Malaysia for the Bachelor
Science (Honours) majoring in Statistics
and graduated in April 1997 before leaving
for Queen’s University, Belfast where he
received the Master’s degree of Science in
Engineering
Computation
with
a
Distinction. He later joined the Intelligent
Systems and Control Research Group at Queen’s University Belfast
in November 1999 where he completed his Doctorate (PhD) on
Online and Constructive Neural Learning Methods. He took
previous employments as a System Engineer and later as a
Lecturer at the Multimedia University Malaysia and also as Senior
Lecturer at the AIMST University. Since November 2006, he has
served as Senior Lecturer at the department of Electrical and
Electronics Engineering, Universiti Teknologi PETRONAS. His
research interests include linear and nonlinear system
identification, model validation and application of intelligent
system in computing and image processing. Dr Vijanth is member
of Institute of Electrical and Electronics Engineering (IEEE) and has
over 40 publications in local and international proceedings and
journal.
Lila Iznita Izhar obtained her BEng in
Electrical and Electronics Engineering from
the University of the Ryukyuus, Japan in
2002. She later earned her MSc in Electrical
and Electronics Engineering from Universiti
Teknologi PETRONAS in 2006. She is
currently a lecturer with Universiti Teknologi
PETRONAS. Her research interests are in
the area of medical image analysis,
computer vision and image processing.
VOLUME Six NUMBER two july - december 2008 PLATFORM
151
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
AUTOMATED PERSONALITY INVENTORY SYSTEM
Wan Fatimah Wan Ahmad*, Aliza Sarlan, Mohd Azizie Sidek
Universiti Teknologi PETRONAS, 31750 Tronoh, Perak Darul Ridzuan, Malaysia
*[email protected]
Abstract
Personality Inventory is a measurement tool to identify the characteristics or traits of an individual. Nowadays,
personality inventory is used globally by many big organisations in order to hire good quality employees.
Therefore, personality inventory is used as part of self-assessment for career planning purposes. The paper
focuses on development of an Automated Personality Inventory System (APIS). The system can be used as a tool
for any organisation especially by the Human Resource Department to assess their (potential) possible employee.
APIS was developed based on the manual Sidek Personality Inventory which determines the characteristics of
individual personality traits. The methodology used in developing the system was the Spiral model. The system
was developed using PHP, MySQL and Apache Web Server. The system may be useful for any organisation and
employee. The system will help an organisation filter the possible candidate as their employee for the qualified
job. The system may also benefit the employee to know their personality type. It is hoped that the system will
help an organisation in choosing the right candidate.
Keyword: personality inventory, organisation, career planning, automated, self-assessment.
Introduction
Nowadays, personality inventories are used globally
by many big organisations in order to hire good
quality employees. A big organisation such as
PETRONAS could use these personality inventories
to filter candidates. Personality inventory could be
used as part of a self-assessment for career planning
purposes. The test results can be enormously helpful
when determining the kind of career a candidate
might like to pursue. Frieswick [1] highlighted that
as employers start to hire again, they are increasingly
taking steps to ensure that the hires they make are a
good fit – not only with the job description but also
with the people with whom they will be working.
Therefore, understanding the candidate’s personality
type will also improve their chance of being happy in
their work. This will decrease the amount of turnover
in an organisation.
Personality tests or inventories are self-reporting
measures of what might be called traits, temperaments,
or dispositions [2]. Different types of tests are used
for different purposes and personality tests are lately
being used more frequently. It is difficult to pinpoint
which tests are more efficient since it is a very
subjective area, however, if used correctly personality
tests can be very effective tools [3]. Research has also
shown that personality can be used as a predictor [4].
In order to create the “perfect” user experience,
designers should also apply a suitable graphical
user interface. Diversity centred approach was
proposed [5] where the design focused more on
groups of users, or even on the characteristics of
individual users. In another words, designers need
to understand individual user characteristics, such as
gender, disabilities or personalities, and relate to user
preferences for interface properties such as structure,
This paper was presented at the International Conference on Science & Technology: Applications in Industry and Education, Penang,
12 - 13 December 2008
152
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
layout or shape. Designers should also segment
their user population according to these designrelevant user characteristics and develop customised
user interfaces. Therefore, successful customisation
would probably improve the usability, and the usage
experience itself. Users would no longer have to
accept the traditional uniform design that only targets
the average user. Instead, they would have a product
tailored to their needs which allows them to express
their individuality.
The goal of the paper is to develop a personality
inventory system with a suitable user interface. This
paper presents an Automated Personality Inventory
System (APIS) which can be used as a tool for any
organisation especially for the Human Resource
Department and to research on the suitable user
interface in the development. APIS is developed
based on the manual Sidek Personality Inventory (IPS).
Currently, IPS personality test has to be done manually
where an individual takes a pencil and paper test.
Related work
A number of personality scales or inventories have
been implemented and made available online. A
majority of them focused on single personality
constructs rather than broad inventories. However,
the evaluation of an online personality test was not
made available [6].
The Hogan Personality Inventory (HPI) provides the
industry a standard for measuring normal personality
[7]. HPI predicts employee performance and helps
companies reduce turnover, absenteeism, shrinkage,
and poor customer service. HPI contains seven primary
scales, six occupational scales, and one validity scale.
The scales are shown in Table 1.
Another online personality inventory that is available
is Jung Typology Test [8]. The indicators used in the
test are divided into four areas: extraversion (E) or
introversion (I); sensing (S) or intuition (N); thinking (T)
Table 1. Hogan personality inventory scales
Primary scales
Adjustment
confidence, self-esteem, and composure under pressure.
Ambition
initiative, competitiveness, and leadership potential.
Sociability
extraversion, gregariousness, and a need for social interaction.
Likeability
warmth, charm, and the ability to maintain relationships.
Prudence
responsibility, self-control, and conscientiousness.
Intellectance
imagination, curiosity, and creative potential.
Learning Approach
the degree to which a person is achievement-oriented and stays up-to-date on
business and technical matters.
Occupational Scales
Service Orientation
being attentive, pleasant, and courteous to clients and customers.
Stress Tolerance
being able to handle stress; low scores are associated with absenteeism and health
problems.
Reliability
integrity (high scores) and organizational delinquency (low scores).
Clerical Potential
the ability to follow directions, pay attention to details, and communicate clearly.
Sales Potential
energy, social skill, and the ability to solve problems for clients.
Managerial Potential
leadership ability, planning, and decision-making skills.
VOLUME Six NUMBER two july - december 2008 PLATFORM
153
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
or feeling (F); and judging (J) or perceiving (P). Results
are usually expressed as four letters indicating the
respective preferences: an example would be ENTP.
A Need for a Personalised Inventory
System
Organisations want the best people to work for them.
To achieve this objective they cannot rely only on the
applicant’s resume. Sometimes people with good
and tempting resume might not be as good as they
claimed. Applicants could cheat on their resume to
make it look impressive in the eye of the organisation
they want to work in. This has become a problem
to many organisations as they have to spend a lot of
time, money and effort in reviewing and screening
all applications to get the best applicant to fill up the
vacant posts. This is where personality inventory
comes in handy because it helps employers analyse
and filter out persons best suitable for the job.
However, good and reliable personality inventories
are hard to find and design. Furthermore, most
personality inventory tests had to be done manually
with the assistance of psychologists for the best
outcome. This setting is impractical as the organisation
has to hire a psychologist for recruitment purposes
only. Therefore, automated personality inventories
should be developed to counter this problem.
APIS development
APIS is developed with the aim in developing a
tool to aid the human resource department of the
organisation in identifying and recruiting a right
person for a right job function. The system serves the
following objectives:
• To effectively filter suitable applicants for any
vacant position in the organisation.
• To easily identify an individual’s suitability for the
job based on their personality analysis.
• Assist individuals to know their personality types.
154
Table 2. Sidek personality inventory traits
Aggressive
Endurance
Analytical
Achievement
Autonomy
Controlling
Depending
Helping
Extrovert
Honesty
Introvert
Self-critics
Intellectual
Supportive
Variety
Structure
APIS is a web-based system that was developed to
automate the conventional personality inventories
system based on Sidek Personality Inventory
(IPS) developed by Associate Professor Dr. Sidek
Mohd Noah [9]. IPS is one of the most successful
personality inventories that could determine the kind
of personality an individual has and also suggests
suitable jobs. There are 16 personality traits that were
used in IPS as shown in Table 2.
The system was developed using Open Source
technology, which consisted of Open Source web
programming language (PHP) or JavaScript, Open
Source database (MySQL), and also Open Source
Web Server (Apache). The project involved Associate
Professor Dr. Sidek Mohd Noah as the client for
the system. The system is developed based on
the spiral development model. Using the model,
user’s requirements and feedback were acquired
throughout the development period as it is crucial to
understand the concept and how it works so that the
system will provide the user with an accurate output
or information.
System architecture
APIS adopted the combination of data-centred and
client-server architectural model. A client-server
system model is organised as a set of services and
associated server(s) and clients that access and use
the services. The server itself or one of the server(s)
contains a database. Figure 1 illustrates the system
architecture.
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
CLIENT
LOGIN page
User
User
User
Internet
Authorized?
New User
Registration
APIS
Selection Page
Login
IPS Questionnaires
Personality Profile
System Page 1
System Page 2
System Page 3
DBMS
SERVER
APIS
Database
Result Page
Figure 1. System architecture
Figure 2. Navigation map
Figure 2 shows the navigation map of the APIS website.
Once a user logs in to the system, the user can choose
which personality inventory system to use from a list
provided. After completing the test, the system will
provide the user with the results page which could be
saved or printed. A user can obtain a display of their
results by logging in back to the system.
simpler, yet custom tailored to users’ specific needs.
In developing APIS, the IPS was used to test the
individual’s personality by posing 160 questions
in a questionnaire form. Each page contained ten
questions, so there were 10 pages for a user to go
through when taking the test. Figure 4 shows the
questionnaires page.
APIS prototype
The prototype is developed to serve two different
types of users namely i) Human Resource staff and ii)
candidates. The main page is a point of access for all
users of the system. Currently, APIS is developed in
Bahasa Malaysia. Figure 3 shows the main page of the
system. According to [10], systems can be designed
Figure 3. Main page
After a user finishes the test, a Personality Profile
that shows the score, percentages and analysis for
each trait will be generated in table form as shown
in Figure 5. Apart from this table, there will also be a
graph and description for the user to see their scores
and personality analysis result as shown in Figure 6.
Figure 4. Questionnaire page
VOLUME Six NUMBER two july - december 2008 PLATFORM
155
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
Figure 5. Personality profile result (table form)
Figure 6. Personality profile result
Functional Testing
A functional testing was conducted with the expert
user, that is, Dr. Sidek. The aim was to validate the
operation of the developed system with respect
to the system’s functional specification and to get
feedback from him regarding the prototype. Overall,
he was satisfied with the system since the functional
requirements were met. With the developed system,
a user does not have to key in the data manually
anymore. The system will be able to automate and
156
present the score, percentages and analysis for each
trait of personality.
Conclusion
The paper presentd an automated personality
inventory system (APIS) which could be very
beneficial for any organisation especially the Human
Resource Department in order to hire quality workers.
Understanding the candidate personality type will
improve their chance of being happy at their work.
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
This will decrease the amount of turnover in an
organisation. Personality is also important since once
people know the type of person they are, they can
learn about how to communicate better, work better
and also change weaknesses into strengths. APIS has
adopted 16 personality traits compared to the others
available online which were only based on 5 traits. It
is hoped that the system will help improve the quality
of employees hired. APIS could also be used by any
individual who wants to know his/her personality
type. Future work could include conversion of the
system to other languages, testing the reliability of
the system and making it accessible to all.
Wan Fatimah Wan Ahmad received her BA
and MA degrees in Mathematics from
California State University, Long Beach,
California USA in 1985 and 1987. She also
obtained Diploma in Education from
Universiti Sains Malaysia in 1992. She
completed her PhD in Information Systems
from Universiti Kebangsaan Malaysia in
2004. She is currently a senior lecturer at
Computer & Information Sciences Department of Universiti
Teknologi PETRONAS, Malaysia. She was a lecturer at Universiti
Sains Malaysia, Tronoh before joining UTP. Her main interests are
in the areas of mathematics education, educational technology,
human computer interaction and multimedia.
References
[1]
Frieswick, K. 2005. Magazine of Senior Executive. Available
online http://www.boston.com/news
[2]
Shaffer, D. J. and Schimdt, R. A. 1999. Personality Testing in
employment. Available online http://pview.findlaw.com/
[3]
Collins, M. 2004. Making the most of Personality Tests.
Available online http://www.canadaone.com/
[4]
Woszczynski, A. B., Guthrie, T. C. and Shade, S. 2005.
Personality and Programming. Journal of Information
Systems Education 16(3), pp. 293-300
[5]
Saati, B., Salem, M. and Brinkman, W. 2005. Towards
customised user interface skins: investigating user
personality and skin colour. Available online. http://mmi.
tudelft.nl/~willem-paul/HCI2005
[6]
Buchanan, T., Johnson, J. A., Goldberg, L. R. 2005. European
Journal of Psychological Assessment, 21(2) pp 116-128
[7]
http://www.hoganassessment.com/_HoganWeb/
[8]
Shaunak, A. 2005. Personality Type – Jung Typology
Test.
Available
online.
http://shaunak.wordpress.
com/2005/10/14/personality-type-jung-typology-test/
[9]
Sidek Mohd Noah, 2008. available online http://www.educ.
upm.edu.my/directori/sidek.htm
[10] Fuchs, R. (2001). Personality traits and their impact on
graphical user interface design: lessons learned from the
design of a real estate website. In Proceedings of the 2nd
Workshop on Attitude, Personality and Emotions in UserAdapted Interaction.
VOLUME Six NUMBER two july - december 2008 PLATFORM
157
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
A Fuzzy Neural Based
Data Classification System
Yong Suet Peng* and Luong Trung Tuan
Universiti Teknologi PETRONAS, 31750 Tronoh, Perak Darul Ridzuan, Malaysia
*[email protected]
Abstract
Data mining has emerged to be a very important research area that helps organisations make good use of the
tremendous amount of data they have. In data classification tasks, fuzzy systems lack the ability to learn and
cannot adjust themselves to a new environment. On the other hand, neural networks can learn, but they are
opaque to the user. This paper presents a hybrid system to perform classification tasks. The main work of this
paper includes generating a set of weighted fuzzy production rules, mapping it into a min-max neural network;
re-deriving the back propagation algorithm for the proposed min-max neural network; and performing data
classification. The iris and credit card datasets were used to evaluate the system’s accuracy and interpretability.
The algorithm was found to have improved the fuzzy classifier.
Keywords: data classification, fuzzy neural
INTRODUCTION
Data mining refers to the discovery step in knowledge
discovery in databases. Its functionalities are used
to specify the kind of patterns to be found in data
mining tasks. In general, data mining tasks can be
classified into 2 categories: descriptive and predictive.
Descriptive mining tasks characterise the general
properties of the data whereas predictive mining
tasks perform inferences on the current data in order
to make predictions [1].
Many algorithms have been proposed and developed
in the data mining field [2][3][4][5]. There are several
challenges that data mining algorithms must satisfy in
performing either descriptive or predictive tasks. The
criteria for evaluating data mining algorithms include:
accuracy, scalable, interpretable, versatile and fast [6]
[7][8].
Although human experts have played an important
role in the development of conventional fuzzy systems,
automatically generating fuzzy rules from data is very
helpful when human experts are not available and
may even provide information not previously known
by experts. Several approaches have been devised
to develop data-driven learning for fuzzy rule based
systems. They involve the use of a method that
automatically generates membership functions of
fuzzy rule structures of both from training data. Chen
[9] proposed a method for generating fuzzy rules for
classification problems which uses fuzzy subsethood
values for generating fuzzy rules.
In Chen’s research [9], he defined fuzzy partitions for
the input variables and output variables according
to the type of data and the nature of classification
problems, then transform crisp value into fuzzy input
value and generate a set of fuzzy rules based on a
This paper was presented at the The 2006 International Conference on Data Mining (DMIN’06),
26-29 June 2006, Las Vegas
158
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
fuzzy linguistics model in order to apply the rule sets
for classification. Regarding the relative weight of the
linguistic term, Khairul [10] proposed the weighted
subsethood based algorithm for data classification.
Weighted subsethood based algorithm is the use
of subsethood values as relative weights over the
significance of different conditional attributes. Fuzzy
systems lack the ability to learn and cannot adjust
themselves to a new environment [15].
our approach
A hybrid system of a neural network with fuzzy system
to data classification was used in this project. Figure 1
shows the block diagram of our developed system.
The system was a supervised learning system. The
training dataset was fed to train the system to
generate the weighted fuzzy production rules (WFPR),
and then these weighted fuzzy production rules were
trained by a min-max neural network to achieve a
better classification accuracy. The system produced a
new set of weighted fuzzy production rule which was
fed into the analysing data into the system and was
applied into a new set of WFPR to classify data.
for every candidate’s input were defined; this step
depended on the target classification class and the
data type of linguistic variable [16].
If the data type of input attribute is nominal, so the
available category will become the linguistic term and
the number of linguistic term is independent with the
number of target class. For example, if the attribute is
gender, which is a nominal data type, then there are
only 2 linguistic terms male and female. In the case
of crisp membership values, a boy will have {(male, 1),
(female, 0)}.
If the data type of the input attribute is continuous
numerical, then the number of linguistic terms equal
to the number of target classes, and the membership
function is trapezoid. For example, with the iris data
set based on the petal length, petal width, sepal
length and sepal width, classification of the iris flower
was into setosa, versicolor, virginica; which meant 3
target classes, and thus 3 linguistic terms for all the
input attributes as shown in Table I. Thus, the value
range of the input attribute value had to be obtained
to define the trapezoid membership.
A. Weighted fuzzy production rules induction
[11][12][13]
The iris data set was used to illustrate the paper
work.
With application on iris data set, the trapezoidal
membership function was used. The iris data set has
4 attributes which are petal length, petal width, sepal
length and sepal width. The iris flower was classified
into setosa, versicolor, virginica.
1) Define input fuzzy member class
2) Calculate the subsethood value
Data classification tasks work on several candidates
were input to classify the observed object into
predetermined classes. In order to generate fuzzy
rules for data classification the linguistics variables
After defining the fuzzy member class, the subsethood
value of each linguistic term were calculated over the
classification class. Table 2 shows the subsethood
values for the iris dataset.
Table 1 Linguistic term for iris dataset
Training
Dataset
Analyzing
Data
Preprocessing
Subsystem
Fuzzy Neural
Network
Fuzzy logic
Inference
Classification
Output & Fuzzy
Rules
Figure 1. Block Diagram of the Fuzzy Neural System
Attributes
Linguistic term
Sepal Length
Small
Average
Large
Sepal Width
Small
Average
Large
Petal Length
Small
Average
Large
Petal Width
Small
Average
Large
VOLUME Six NUMBER two july - december 2008 PLATFORM
159
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
Table 2. Subsethood values for iris dataset
setosa
versicolor
virginica
Small
S(set,S_SL)
S(ver,S_SL)
S(vir,S_SL)
Average
S(set,A_SL)
S(ver,A_SL)
S(vir,A_SL)
Large
S(set,L_SL)
S(ver,L_SL)
S(vir,L_SL)
Small
S(set,S_SW)
S(ver,S_SW)
S(vir,S_SW)
Average
S(set,A_SW)
S(ver,A_SW)
S(vir,A_SW)
Large
S(set,L_SW)
S(ver,L_SW)
S(vir,L_SW)
Small
S(set,S_PL)
S(ver,S_PL)
S(vir,S_PL)
Average
S(set,A_PL)
S(ver,A_PL)
S(vir,A_PL)
Large
S(set,L_PL)
S(ver,L_PL)
S(vir,L_PL)
Petal
Small
S(set,S_PW)
S(ver,S_PW)
S(vir,S_PW)
width
Average
S(set,A_PW)
S(ver,A_PW)
S(vir,A_PW)
Large
S(set,L_PW)
S(ver,L_PW)
S(vir,L_PW)
setosa
versicolor
virginica
Small
W(set,S_SL)
W(ver,S_SL)
W(vir,S_SL)
Average
W(set,A_SL)
W(ver,A_SL)
W(vir,A_SL)
Large
W(set,L_SL)
W(ver,L_SL)
W(vir,L_SL)
Small
W(set,S_SW)
W(ver,S_SW)
W(vir,S_SW)
Average
W(set,A_SW)
W(ver,A_SW)
W(vir,A_SW)
Large
W(set,L_SW)
W(ver,L_SW)
W(vir,L_SW)
Small
W(set,S_PL)
W(ver,S_PL)
W(vir,S_PL)
Average
W(set,A_PL)
W(ver,A_PL)
W(vir,A_PL)
Large
W(set,L_PL)
W(ver,L_PL)
W(vir,L_PL)
Small
W(set,S_PW)
W(ver,S_PW)
W(vir,S_PW)
Average
W(set,A_PW)
W(ver,A_PW)
W(vir,A_PW)
Large
W(set,L_PW)
W(ver,L_PW)
W(vir,L_PW)
Sepal length
Sepal width
Petal length
Table 3: Weighted subsethood values for iris dataset
Sepal length
Sepal width
Petal length
Petal width
Here, S(set, S_SL) is the subsethood of Small_
SepalLength with regards to setosa and the rest
follow.
4) Weighted fuzzy rule generation
3) Calculate the weighted subsethood value
• If Sepal length is [W(set,S _ SL) x Small
After getting the weighted subsethood values, the set
of weighted fuzzy production rule was generated.
OR W(set,A _ SL) x Average OR W(set,L _ SL) x
Based on Table 2, the weighted subsethood value
was calculated. As in Table 3, W(set, S_SL) refers to
the weighted subsethood of Small_SepalLength with
regards to setosa and the rest follow.
160
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Large] AND Sepal width is [W(set,S _ SW) x
Small OR W(set,A _ SW) x Average OR W(set,L _
SW) x Large] AND Petal Length is [W(set,S _
PL) x Small OR W(set,A _ PL) x Average OR
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
•
W(set,L _ PL) x Large] AND Petal width is
If
[W(set,S _ PW)
x
Small
OR
W(set,A _ PW)
W’(set,SL) x {Sepal length is [W’(set,S _ SL)
x
x Small OR W’(set,A _ SL) x Average OR W’(set,L _
Average OR W(set,L _ PW) x Large] Then Class
SL) x Large]} AND W’(set,SW) x {Sepal width is
is SETOSA
[W’(set,S _ SW) x Small OR W’(set,A _ SW) x Average
If Sepal length is [W(ver,S _ SL) x Small
OR W’(set,L _ SW) x Large]} AND W’(set,Pl) x {Petal
OR W(ver,A _ SL) x Average OR W(ver,L _ SL) x
Length is [W’(set,S _ PL) x Small OR W’(set,A _
Large] AND Sepal width is [W(ver,S _ SW) x
PL) x Average OR W’(set,L _ PL) x Large]} AND
Small OR W(ver,A _ SW) x Average OR W(ver,L _
W’(set,PW) x {Petal width is [W’(set,S _ PW) x
SW) x
Small OR W’(set,A _ PW) x Average OR W’(set,L _
Large] AND Petal Length is [W(ver,S _
PL) x Small OR W(ver,A _ PL) x Average OR
PW) x Large]} Then Class is SETOSA;
W(ver,L _ PL) x Large] AND Petal width is
[W(ver,S _ PW)
x
Small
OR
W(ver,A _ PW)
x
Average OR W(ver,L _ PW) x Large] Then Class
is VERSICOLOR
• If Sepal length is [W(vir,S _ SL) x Small
Where W’(set,SL) x W’(set,S _ SL) = W(set,S _ SL);
initially for neural network weight W’(set,S _ SL) =
W(set,S _ SL) and W’(set,SL) =1, similarly for the
other weights.
OR W(vir,A _ SL) x Average OR W(vir,L _ SL) x
Large] AND Sepal width is [W(vir,S _ SW) x
Small OR W(vir,A _ SW) x Average OR W(vir,L _
SW) x Large] AND Petal Length is [W(vir,S _
PL) x Small OR W(vir,A _ PL) x Average OR
The rule was a logic AND operation of sets of OR logic
operation. After the rules were re-written, a set of
sub-rules were obtained which has OR operation and
AND operation rules.
W(vir,L _ PL) x Large] AND Petal width is
x
Sub rule 1: if Sepal length is [W’(set,S _ SL) x
Average OR W(vir,L _ PW) x Large] Then Class
Small OR W’(set,A _ SL) x Average OR W’(set,L _
is VIRGINICA
SL) x Large] Then Class is SETOSA valued at
[W(vir,S _ PW)
x
Small
OR
W(vir,A _ PW)
A1
B. Fuzzy Neural network subsystem
Sub rule 2: if Sepal width is [W’(set,S _ SW) x
The task of the fuzzy neural network subsystem
was to train the weighted fuzzy production rules by
modifying the weight so that the rule can classify data
with higher accuracy. Through the literature review
the fuzzy neural network generally has the input
membership layer, fuzzy rule layer and output layer.
The activation function for the input membership
layer is the trapezoid membership function; the
activation for the fuzzy rule layer is the fuzzy rule, and
the activate function for the output layer is the logic
operation function.
Small OR W’(set,A _ SW) x Average OR W’(set,L _
SW) x Large] Then Class is SETOSA valued at
A2
Sub rule 3: if Petal Length is [W’(set,S _ PL) x
Small OR W’(set,A _ PL) x Average OR W’(set,L _
PL) x Large] Then Class is SETOSA valued at
A3
Sub rule 4: if Petal width is [W’(set,S _ PW) x
Small OR W’(set,A _ PW) x Average OR W’(set,L _
PW) x Large] Then Class is SETOSA value at A4
The generated weighted production fuzzy rules were
the combination of “And” and “Or” operations, which
cannot be used to activate the neuron. Thus, the
rule was separated into two sub-rules; one sub-rule
performs the “OR” operation and another sub-rule
performs the “AND” operation. The initial rule can be
rewritten as follows:
Followed by: If W’(set,SL) x A1 AND W’(set,SW) x
A2 AND W’(set,Pl) x A3 AND W’(set,PW) x A4 then
Class is Setosa
Simplification of the the rule helped in the modification
of the fuzzy rule layer in the fuzzy neural network
VOLUME Six NUMBER two july - december 2008 PLATFORM
161
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
into two sub-layers: AND (Min) layer and OR (Max)
Layer.
5) Fuzzy neural network design
Figure 2 shows the fuzzy neural network architecture
for the system.
Layer 1 Input layer:
Each neuron in this layer transmits external crisp
signals directly to the next layer
Y(1)i = X(1)i
Where X(1)i is the input, Y(1)i is the output of neuron
i in layer 1
Layer 2 input membership layer:
Layer 5 defuzzification
The output was classified. For the classification
purpose, a minor modification was made for the
computed output of network
1, if y i(4) = Max i =1..n y i(4)
Yi (5) = 
(4)
(4) 0, if y i ≠ Max i =1..n y i
X(2)j = Y(1)i
Layer 1
Crisp
Input
X1
These layers perform the AND and OR operation in
the weighted fuzzy rule. The activation for the AND
layer was re-derived which is the min operation, and
the activation function for the OR layer is the max
operation.
Layer 4
Min
Layer
Layer 5
Output
maxW11
R11
minW11
R21
A2
A3
R12
maxW32
O1
minW31
minW22
B1
R13
X2
(7)
Where maxC is number of neuron in layer 3; maxWi,j
is the weight of connection from neuron i in layer 2 to
neuron j in layer 3
(10)
Layer 3
Max
layer
maxW12
Layer 3 and Layer 4 are the fuzzy rule layers:
162
Layer 2
Input
membership
functions
A1
Y(2)j = μ(X(2)j)
C
( 2)
Y j( 3) = ∨ imax
), = 0 (max Wi , j ∗ Yi
1 OC (5)
∑ (Yk − y k ) 2 2 k =1
Where OC is number of output neurons, yk is the
desired output for neuron k and Yk(5) is the actual
output of neuron k. It could be seen that the error
E(p) is the function with respect to minW, maxW. The
main objective is to adjust these weights such that
error function reaches minimum or is less than a given
threshold.
The output of neuron in this layer is the membership
value of crisp input value
(9)
For 1 iteration the error was
E ( p) =
Neurons in this layer represent fuzzy sets used in the
antecedents of fuzzy rules. A fuzzification neuron
receives a crisp input and determines the degree to
which this input belongs to the neuron’s fuzzy set.
The activation function is the membership function.
The input is the output of crisp input layer
C
( 3)
Y j( 4 ) = ∧ imin
), = 0 (min Wi , j ∗ Yi
(8)
Where minC is number of neurons in layer 4; minWi,j
is the weight of connection from neuron i in layer 3 to
neuron j in layer 4
R22
B2
B3
R14
minW42
Figure 2. Fuzzy neural network architecture
PLATFORM VOLUME Six NUMBER TWO july - december 2008
O2
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
6) Learning in fuzzy neural network
After training, new set of weights were obtained, and
the trained rules became:
For the learning phase, we apply the back propagation
equations[14][15]. According to the principle of
gradient descent, the back propagation equations
can be written as
If minW(set,SL) x {Sepal length is [maxW(set,S _
min Wi , j ( p + 1) = min Wi , j ( p ) − α ∗
maxW(set,L _ SL) x Large]}
AND minW(set,SW) x {Sepal width is [maxW(set,S _
SW) x Small OR maxW(set,A _ SW) x Average OR
∂E ( p )
∂ min Wi , j
max Wi , j ( p + 1) = max Wi , j ( p ) − α ∗
SL) x Small OR maxW(set,A _ SL) x Average OR
maxW(set,L _ SW) x Large]}
∂E ( p )
∂ max Wi , j
AND minW(set,Pl) x {Petal Length is [maxW(set,S _
PL) x Small OR maxW(set,A _ PL) x Average OR
maxW(set,L _ PL) x Large]}
where α is the learning rate.
AND minW(set,PW) x {Petal width is [maxW(set,S _
PW) x Small OR maxW(set,A _ PW) x Average OR
By eliminating the derivative process, the following
results were obtained:

(Y j( 4 ) − Y j ) ∗ Yi (3) , C1
∂E ( p )
=  ( 4)
C
( 3)
∂ min Wi , j (Y j − Y j ) ∗ Yi (3) * ∧ min
ki = 0 , k ≠ i (min W k , j ∗ Yk ), C 2
C
( 3)
C1 : min Wi , j ∗ Yi (3) is ∧ min
k = 0 (min W k , j ∗ Yk )
C
( 3)
C 2 : min Wi , j ∗ Yi (3) is not ∧ min
k = 0 (min W k , j ∗ Yk )
∂E ( p )
∂ max Wi , j
maxW(set,L _ PW) x Large]}
Then Class is SETOSA
The rules were simplified as follows:
If Sepal length is [newW(set,S _ SL) x Small OR
newW(set,A _ SL) x Average OR newW(set,L _ SL) x
Large]
 (Yu( 4 ) − Yu ) ∗ min W j ,u ∗ Yi 2 , if C1


( 3)
C
 (Yu( 4 ) − Yu ) ∗ ∧ kimin
)
= 0 , k ≠ j (min Wk ,u ∗ Yk

2
 ∗ min W j ,u ∗ Yi , if C 2


=
(Yu( 4 ) − Yu ) ∗ min W j ,u


∗ max Wi , j ∗ Yi ( 2 ) ∗ Yi 2 , if C 3


(Y ( 4 ) − Y ) ∗ ∧ min C (min W ∗ Y ( 3) )
u
ki = 0 , k ≠ j
k ,u
k
 u
∗ min W j ,u ∗ max Wi , j ∗ Yi ( 2 ) ∗ Yi 2 , if C 4

AND Sepal width is [newW(set,S _ SW) x Small OR
newW(set,A _ SW) x Average OR newW(set,L _ SW) x
Large]
AND Petal Length is [newW(set,S _ PL) x Small
OR newW(set,A _ PL) x Average OR newW(set,L _ PL)
x Large]
AND Petal width is [newW(set,S _ PW) x Small OR
newW(set,A _ PW) x Average OR newW(set,L _ PW) x
Large]
Then Class is SETOSA
Here,
( 3)
j
C1 : min W j ,u ∗ Y is ∧
min C
k =0
( 3)
k
(min Wk ,u ∗ Y
)
C
( 2)
and max Wi , j ∗ Yi ( 2 ) is ∨ max
k = 0 (max Wk , j ∗ Yki )
C
( 3)
C 2 : min W j ,u ∗ Y j( 3) is not ∧ min
k = 0 (min W k ,u ∗ Yk )
C
and max Wi , j ∗ Yi ( 2 ) is ∨ max
(max Wk , j ∗ Yki( 2 ) )
k =0
C
( 3)
C 3 : min W j ,u ∗ Y j(3) is ∧ min
k = 0 (min W k ,u ∗ Yk )
C
and max Wi , j ∗ Yi ( 2 ) is not ∨ max
(max Wk , j ∗ Yki( 2 ) )
k =0
C
( 3)
C 4 : min W j ,u ∗ Y j(3) is not ∧ min
k = 0 (min W k ,u ∗ Yk )
C
and max Wi , j ∗ Yi ( 2 ) is not ∨ max
(max Wk , j ∗ Yki( 2 ) )
k =0
newW(set,S _ SL)
=
maxW(set,S _ SL)
x
minW(set,SL)
Results and discussions
With the iris data set, the entire dataset was labeled
from 1 to 150, divided equally into two sub-datasets:
IP1 and IP2; IP1 consisted of the odd numbered
objects and IP2 consisted of the even numbered
objects. Then the sub-dataset was fed to train the fuzzy
neural system and to analyse data. For evaluation, the
IP1 was used to train the system, and then the IP2 and
the whole dataset was analysed; and the IP2 was also
used to train the system then the IP1 and the whole
dataset was analysed. With this, the accuracy of results
achieved are shown in Table 4.
VOLUME Six NUMBER two july - december 2008 PLATFORM
163
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
Table 4. Iris data set result
Training
dataset
Training
Accuracy
Testing
dataset
SBA Fuzzy
System
WSBA Fuzzy
System
Fuzzy Neural
System
IP1
100%
IP2
80%
93.33%
94.67%
Whole
78.69%
94.67%
96.67%
IP1
78.67%
93.33%
93.33%
Whole
78%
93.33%
96.77%
IP2
97.30%
Table 5. Credit Card data set result
Training
dataset
Training
Accuracy
Testing
dataset
SBA Fuzzy
System
WSBA Fuzzy
System
Fuzzy neural
System
Cr1
80%
Cr2
70%
76%
81%
SBA: subsethood based algorithm [10]
WSBA: weighted subsethood based algorithm [10]
For credit card dataset, the dataset was divided into 2
sub-datasets, one was used for training and the other
was used for evaluating the system. 358 samples were
randomly selected from the credit card approval data
set for training set (Cr1) and wthe rest of 332 samples
(Cr2) were used for testing. The achieved training
accuracy was 80% and the evaluation accuracy was
81% as shown in Table 5.
conclusion
Based on the results from Table 4 and Table 5,
the hybrid system with fuzzy neural network has
improved the accuracy of the weighted fuzzy
production rules. For the iris dataset, the improvement
was about 1-2% accuracy increased whereas for
the credit card approval dataset 5% accuracy
improvement was obtained as compared to fuzzy
systems.
This project has demonstrated a fuzzy neural
algorithm for data classification task. Fuzzy logic
and neural networks are complementary tools in
building intelligent systems. Fuzzy systems lack the
ability to learn and cannot adjust themselves to a new
environment. On the other hand, although neural
networks can learn, they are opaque to the user
[15]. Integrated fuzzy neural systems can combine
the parallel computational and learning abilities of
neural networks with the human-like knowledge
representation and explanation abilities of fuzzy
system. As a result neural networks has become more
transparent, while fuzzy systems has become capable
of learning.
With fuzzy neural approach, the initial weight was
taken from the weighted fuzzy rules that had high
accuracy; thus there was no need to randomly
generate the weight and start training from without
knowledge. Besides that, the hybrid system could
generate interpreted rules that were not applicable in
neural networks.
164
Data mining has emerged to be a very important
research area that helps organisations make good use
of the tremendous amount of data they have. With
the combination of many other research disciplines,
data mining turns raw data into useful information
rather than just raw, meaningless data.
In conclusion, the hybrid system could gain high
accuracy for data classification task. Besides that,
it could generate rules that helped to interpret the
output results. Also, the neural network being used
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
was the easiest and most popular network that was
a min-max back propagation network. Therefore, it
makes the algorithm easy to implement. The algorithm
has improved the fuzzy classifier as well as the neural
network training. However, the processing time will
be long if the dataset is huge. Thus, for future work,
further testing needs to be done in order to improve
the design of the algorithm to classify larger datasets.
References
[1]
Margaret H. Dunham, 2003, “Data Mining Introductory and
Advanced Topics”, New Jersey, Prentice Hall
[2]
Mehmeh Katardzic, 2003, “Data Mining: Concepts, Models,
and Algorithm”, John Wiley & Sons
[3]
David Hand, Heikki Mannila, Padhraic Smuth, 2002,
“Principles of Data Mining”, The MIT Press
[4]
Jiawei Han, Micheline Kamber, 2000, “Data Mining: Concepts
and Techniques”, Morgan Kaufmann
[5]
Freitas, A. A., 2002, “Data Mining and Knowledge Discovery
With Evolutional Algorithms”, Neural Computing series
[6]
John W., 2003, “Data Mining: Opportunities and Challenges”,
Idea Group Publishing
[7]
Michalski R. et al., 1998, “Machine Learning and Data Mining:
Methods and Application” John Wiley & Son
[8]
Karuna Pande Joshi, 1997, “Analysis of Data Mining
Algorithms”, <http://userpages.umbc.edu/~kjoshi1/datamine/proj_rpt.htm>. Accessed in January 2005
[9]
Chen S. M., Lee S. H., and Lee C. H., 2001, “A new Method for
generating Fuzzy Rules From numerical data for handling
Classification Problems”, Applied Artificial Intelligence, vol.
15, pp 645-664.
[15] Micheal Negnevitsky, 2005, “Artificial Intelligence, A guide
to Intelligent Systems”, Second Edition, Addison Wesley
[16] Zadeth L. A., 1988, Fuzzy Logic IEEE – Computer Science,
vol. 21, pp83-93
S. P. Yong obtained her Master in
Information Technology and Bachelor in
Information Technology (Honours) in
Industrial Computing from Universiti
Kebangsaan Malaysia. Her employment
experience includes IBM Malaysia Sdn Bhd,
Informatics Institute Ipoh and Univerisiti
Teknologi PETRONAS.
Her research
interests are in intelligent systems, data
and text mining.
[10] Rasmani K. A., Shen Q., 2003, “Weighted Linguistic modeling
based on fuzzy subsethood values”, IEEE International
Conference on Fuzzy Systems, USA (FUZZ-IEEE 2003)
[11] Castro J. L. and Zurita J. M., 1997, “An inductive Learning
Algorithm in fuzzy Systems”, Fuzzy Sets and Systems, vol.
89, pp 193-203
[12] D. S. Yeung and E. C. C. Tsang, 1997, “Weighted Fuzzy
Production rules”, Fuzzy Sets and Systems, vol. 88 pp229313
[13] Hong T. P. and Lee. C. Y., 1996, “Induction of Fuzzy Rules and
Membership Functions from Training Examples”, Fuzzy sets
and Systems, vol. 84, pp33-47
[14] Eric C. C. Tsang, Daniel So Yeung and Xi-Zhao Wang, 2002,
“Learning Weights of Fuzzy Production Rules by A Max-min
neural network”, Proceedings of IEEE Conference on Systems,
Men and Cybernetics, pp1485-1490, Tuscon Arizona USA.
VOLUME Six NUMBER two july - december 2008 PLATFORM
165
OTHER AREAS
RESEARCH IN EDUCATION:
TAKING SUBJECTIVE BASED RESEARCH SERIOUSLY
ABSTRACT
Sumathi Renganathan and Satirenjit Kaur
Universiti Teknologi PETRONAS, 31750 Tronoh, Perak Darul Ridzuan, Malaysia.
Qualitative based research is often criticized for being bias and unscientific. In an educational institution
dominated by researchers in the scientific discipline where positivism’s methodology reigns supreme, it is
often difficult to convince research colleagues of the true value of research that are subjective. In this paper the
authors seek to investigate how other researchers from the scientific community, namely from the engineering
discipline, who are familiar with scientific based research perceive education research that are subjective and
value-laden. It is also hoped that the findings from this study can highlight that different epistemologies help
produce different knowledge, and that subjective research can produce noteworthy research findings in the
field of education.
Keywords: qualitative, subjective methodology, interpretivism, education research
Introduction
Research is a process of finding solutions to a problem
by conducting a thorough study and analysis of
the situational factors. It provides the needed
information that guides people to make informed
decisions to successfully deal with problems (Sekaran,
2000). This can be done by conducting analysis of
data gathered firsthand (primary data) or data that
are already available (secondary data). The data can
either be quantitative which are gathered through
experiments, structured questions, numerical
information, or qualitative which are gathered using
open-ended questions through interviews or through
observations. However, the idea of research varies in
different disciplines or subject areas.
In an educational institution dominated by
researchers in the scientific discipline where
positivism’s methodology reigns supreme, it is often
difficult to convince research colleagues of the true
value of research that are subjective. In the educational
institution where we carried out this study, research
papers from a particular discipline are often reviewed
by colleagues from some other discipline. Thus, we
believe that there is often major conflict of ideas and
underpinning knowledge on what “research” from
various disciplines should be.
This research is carried out in a technological
university where the majority of lecturers are from
the engineering discipline. In the Engineering
discipline itself there are different departments such
as the Electrical and Electronics, Chemical, Civil,
and Mechanical. In addition, research among these
disciplines may differ, however, majority of these
researchers are from the hard-science and thus we
believe have a positivist view of what a research
should be. We believe that it is much more difficult to
convince engineering colleagues of the true value of
This paper was presented at the ICOSH 07 Conference, Kuala Lumpur,
13 - 15 March 2007
166
PLATFORM VOLUME Six NUMBER TWO july - december 2008
OTHER AREAS
research which falls under the soft-science especially if
there is a conflict in the underlying research paradigm.
There has been ongoing debate in the social sciences
regarding appropriate research methodologies. In
general, there are two main paradigms in research,
known as positivist (usually practiced by researchers
from the hard-science) and non-positivist (typically
practiced by researchers from the soft-science). The
debate about which paradigm needs to be adopted
revolves around the fact that each paradigm has its
strengths and limitations.
The positivist paradigm is also known as traditional,
experimental, empiricist and quantitative. The origins
of this paradigm date back to the nineteenth century
during the period of the French philosopher, August
Comte. There are several other well-known scholars
of this paradigm including Durkheim, Weber and Marx
(Creswell, 2003). This approach argues that the world
is external and objective. The positivists suggest that
there are certain laws which can be found in both
social and natural worlds in a sequence of cause and
effect, and believe in causality in the explanation of
any phenomena. Causality explains the relationship
between cause and effect and integrates these
variables into a theory. In this paradigm, the observer
is considered to be an independent entity. It is
believed that the researcher can remain distant from
the research, in terms of their own values which might
distort their objectivity. Researchers that adhere to this
paradigm operationalise the research concepts to be
measured, and large samples are taken. This research
approach is known as the quantitative approach and
is based on a numerical measurement of a specific
aspect of phenomena. The aim of this approach is to
seek a general description or to test causal hypotheses
(Hollis, 1994).
On the other hand, the non-positivist paradigm
considers the world to be socially constructed
and subjective. This paradigm encompasses the
constructivist approach, interpretative approach,
postmodern perspective and it is commonly known
as the qualitative approach. The constructivist or
interpretivist believe that to understand the world
of meaning one must be able to interpret it. The
inquirer must elucidate the process of meaning and
clarify what and how meanings were embodied in the
language and actions of social actors (Love et al., 2002,
p. 301). Guba and Lincoln (1994) argued that post
modernism suggested that the world was constituted
by “our shared language and we can only know the
world through the particular forms of discourse
that our language(s) can create” (Love et al., 2002, p.
298). Based on these, this paradigm in general, relies
on personal views, experiences and subjectivity in
understanding human actions and behaviour. The
observer is considered part of what is being observed.
Researchers of this paradigm use multiple methods
to establish different views of a phenomena. Smaller
samples are taken to investigate in-depth or over
time, and the research approach used is known as
qualitative.
The purpose of this study
We carried out this study to investigate how other
researchers from the scientific community, namely
from the engineering discipline, who are familiar
with scientific based research perceive education
research that are subjective and value-laden. It is not
our intention to criticize any research paradigm or the
various research methods used for research purposes
because we believe as others (McGrath, 1982; Frankel,
2005), different research problems require different
research methodologies. What we hope, is that this
study can address any gap that might exist in the
engineering lecturers’ understanding of subjective
education research.
In this paper, first, we discuss based on existing
literature, what is the general understanding of
research. Next we give a brief background of this study
followed by a discussion on the findings and analysis.
This is followed by a further discussion on the findings
through answering the two main research questions:
1. How familiar are the researchers from the
engineering discipline, in subjective, qualitative
based education research?
2. What are the engineering researchers’ perceptions
regarding the contribution of qualitative based
research?
VOLUME Six NUMBER two july - december 2008 PLATFORM
167
OTHER AREAS
Finally, we conclude by highlighting the contributions
of this study.
General Understanding of Research
In this paper, our main underlying notion is that of
Hitchcock and Hughes (1995) who suggested that
ontological assumptions give rise to epistemological
assumptions, which give rise to methodological
considerations, which in turn give rise to
instrumentation and data collection. Thus, research is
not merely about research methods:
… research is concerned with understanding of the
world and that this is informed by how we view our
world(s), what we take understanding to be, and
what we see as the purposes of understanding.
Cohen, Manion and Morrison, 2000:1
This paper examines the practice of research through
two significant lenses (Cohen, Manion and Morrison,
2000):
1. scientific
and
positivistic
methodologies
(quantitative approach)
2. naturalistic and interpretive methodologies
(qualitative approach)
These two competing views have been absorbed by
educational research. ‘Positivism strives for objectivity,
measurability, predictability, controllability, patterning
the construction of laws and rules of behaviour,
and the ascription of causality’ (pg28). Positivist
educational researchers ‘concentrate on the repetitive,
predictable and invariant aspects of the person’
whereas interpretivists explore the person’s intention,
individualism and freedom (Cohen, Manion and
Morrison, 2000:18). Interpretivist strive to understand
and interpret the world in terms of its actors’ (pg 28).
As discussed earlier, terms such as hard-science and
soft-science are also often used to refer to positivistic
and interpretive methodologies respectively.
Positivists often argue that interpretivists often do
not use scientific methods, use anecdotal evidence,
or evidence which are often not mathematical
thus, these methods are often considered to be
‘lack of rigour’. To conduct research in any field, it
168
is important to identify the research paradigm of
the researcher. As Guba and Lincon (1994) state, a
research paradigm guides the researcher in choices of
methods as well as epistemologically. Epistemology
is defined as ‘the study of the nature of knowledge
and justification’ (Moser, 1995), which implies that it
“involves both articulating and evaluating conceptions
of knowledge and justification’ (Howe, K.R., 2003:97).
Research paradigms may vary from positivist
(evaluative-deductive) paradigm to interpretivist
(evaluative-interpretivist) paradigm and there are also
combinations of paradigms which give rise to mixed
paradigms (Johl, S.K, Bruce, A. and Binks, M. 2006).
Research paradigms often guide the researchers in
identifying suitable methods for their research and
more often than not many researchers distinguish
their research as either qualitative or quantitative.
Some researchers also use both the quantitative and
qualitative methods and refer it as a mixed-method
research. Although researchers may use various
methods in their research, a positivist epistemological
stance is radically different from an interpretivist.
However, often a positivist paradigm is identified with
quantitative methods and the interpretivist paradigm
is identified with qualitative methods.
The strength of quantitative research approaches, in
the field of education, is that they are able to yield data
that is projectable to a larger population. Furthermore,
because quantitative approaches deal with numbers
and statistics, the data can be effectively translated
into charts and graphs. Thus, quantitative research
methods play important role in policy making in
the field of education. As for qualitative research
approaches, the strength are that they are able to
gain insights of participants’ feelings, perception
and viewpoints. These insights are unobtainable
from quantitative research methods. In the field
of education the findings from qualitative research
can determine “why” and “how” certain educational
policies actually work (or not work) in the real world.
The following is a brief summary of the differences
between qualitative and quantitative research
methods (Burns, 2000):
PLATFORM VOLUME Six NUMBER TWO july - december 2008
OTHER AREAS
Table 1. Summary of qualitative and quantitative research methods
Qualitative
Assumptions •
•
•
•
Quantitative
Reality socially constructed
Variables complex and interwoven, hard
to measure
Events viewed from informant’s
perspectives
Dynamic quality to life
•
•
•
•
Facts and data have objective reality
Variables can be measured and identified
Event’s viewed from outsider’s
perspective
Static reality to life
Purpose
•
•
•
Interpretation
Contextualisation
Understanding of the perspectives of
others
•
•
•
Prediction
Generalisation
Causal explanation
Method
•
•
•
•
•
•
•
Testing and measuring
Commences with hypothesis and theory
Manipulation and control
Deductive and experimental
Statistical analysis
Abstract impersonal write-up
•
•
Data collection using participant
observation, unstructured interviews
Concludes with hypothesis and
grounded theory
Emergence and portrayal
Inductive and naturalistic
Data analysis by themes from informants’
descriptions
Data reported in language of informant
Descriptive write-up
•
•
•
Researcher as instrument
Personal involvement
Empathetic understanding
•
•
•
Researcher applies formal instruments
Detachment
Objective
•
•
•
•
Role of
Researcher
The Study
This study was carried out in a private technological
university which offers undergraduate and
postgraduate degree programmes in Engineering
and Information Technology/Information System (IT/
IS). We used questionnaires to obtain the relevant
information and the participants were lecturers from
the Engineering and the IT/IS discipline. We only
targeted lecturers who possess a PhD as we believe
that these lecturers would practice and understand
research better. The university has a total of 263
lecturers whereby 37% (98 lecturers) possess a PhD.
We distributed the questionnaires during a workshop
that was particularly conducted for identified
lecturers from every department in the university, for
their research experiences or potential capabilities
in research. Thirty-three lecturers (33%) completed
and returned the questionnaires. The questionnaires
contained 28 statements related to quantitative and
qualitative based research. However, each statement
required the lecturers to rate on a 7-point Likert scale,
thus, there is no right or wrong answers but based
on their ratings it will be possible to gauge lecturers’
familiarity and perception of qualitative based
research.
Findings and Analysis of Data
Although the questionnaire contained 28 statements,
these statements were preceded with the following
question: How familiar are you with qualitative based
research?
VOLUME Six NUMBER two july - december 2008 PLATFORM
169
OTHER AREAS
Not at all
familiar
1
Very
familiar
2
3
4
5
6
7
For this question, we categorised lecturers whose
responses ranged between 1 and 3 as indicating that
they are not familiar and responses that ranged
between 5 and 7 as familiar while the mid-point 4 as
those who believe that they have some knowledge
regarding qualitative research. Surprisingly 27%
(9 lecturers) choose not to answer this question.
Furthermore, equal number of lecturers responded for
each category, 8 lecturers (24%) responded as being
not familiar, another 8 as being familiar and the rest
(again another 8 lecturers) responded as having some
knowledge of qualitative research. It is also important
to note that 4 of the lecturers in this study stated that
they are not at all familiar with qualitative based
research and thus did not complete the questionnaire.
Based on these, we decided that it was important
to analyse the lecturers’ responses for each of the
statements in the questionnaire. However, we have
analysed the lecturers’ responses to the statements
under the following headings:
•
•
•
•
•
General perception of “research”
Research paradigm
Research methods
Researcher’s role
Style of presentation
General perception of “research”
We included 6 statements in the questionnaires to
gauge the lecturers’ general perception regarding
research, especially qualitative research:
Each statement required the lecturers to choose from
a 7-point Lickert scale whereby 1 indicates strongly
disagree, 4 as neutral and 7 as strongly agree.
Based on the mean scores shown in Table 2, the
findings indicate that the lecturers in this study are
not that familiar with qualitative research. In fact, the
lower mean scores (3.97 and 3.94) is for Statements 1
and 19 which indicate that the lecturers disagree with
these statements. This implies that they believe that
all researchers have the same research paradigm and
thus, regard research from different disciplines as the
same. Furthermore, the mean scores for the lecturers’
responses for the other statements (Statements
2, 22, 23, 24) ranged from 4.5 to 4.8, which indicate
the lecturers uncertainty on their understanding of
qualitative based research.
Research Paradigm
Although we perceive the lecturers from the
engineering discipline as positivists, we wanted to
know their perception regarding qualitative based
research. Our statements in the questionnaire did
not use explicit terms for research paradigms such as
positivist or interpretivist , in case the lecturers were
Table 2. Lecturers’ general perception of “research”
No.
Statements
Mean Scores
1
The idea of “research” varies in different disciplines or subject areas
3.97
2
In qualitative based research, meanings are discovered not created
4.12
19
Researchers from different disciplines think about “research” differently
3.94
22
Qualitative research seeks to illuminate, understand and extrapolate what is being
researched.
4.82
23
Qualitative research is concerned with meanings and personal experience of individuals,
group and sub-culture.
4.70
24
Research is about making sense of chaos and translating it into culturally accepted
explanations.
4.50
170
PLATFORM VOLUME Six NUMBER TWO july - december 2008
OTHER AREAS
Table 3. Statements to gauge positivist influence on qualitative based research
No.
Statements
Mean Scores
3
A good qualitative research must be reliable, valid and objective.
5.67
5
In qualitative research, the truth should b established logically or supported by
empirical evidence.
5.15
6
The research must be carried out systematically so that it can be replicated by other
researchers.
5.39
7
Qualitative research seeks causal determination, prediction and generalisation of
findings.
4.39
8
The quality of the research is determined by the validity of the research.
5.12
10
Doing research means getting the right answer or the true picture of what was
being researched.
4.48
11
A qualitative research believes in a true reality and the task of research is to find it.
4.42
14
The research must be free of biases.
5.00
16
Rules for qualitative based research are objectivity, consistency and rationality.
5.09
18
The researcher believes that there is a reality out there that can be objectively
measured and found through research.
4.36
20
Qualitative findings are not generalizable.
3.21
27
There must be real or tangible conclusion in a research.
4.70
28
Research must be able to produce findings that are conclusive.
4.61
not familiar with such terms. However, based on their
responses it was possible to identify if the lecturers’
perceptions regarding qualitative based research are
influenced by their positivist research paradigm.
As stated earlier, to conduct research in any field,
it is important to identify the researcher’s research
paradigm. Furthermore, a research paradigm guides
the researcher in choices of methods as well as
epistemologically (Guba and Lincon, 1994). Thus,
in this study we included 13 statements (see Table
3) to gauge whether the lecturers were influenced
by their positivist research paradigm in determining
qualitative based research:
The mean scores for the responses from the lecturers
for these questions ranged from 4.36 (for Statement
8) to 5.67 (for Statement 3). The only exception
is for Statement 20, whereby researchers from an
interpretivist paradigm would have had scores ranging
from 5 to 7. However, a mean score of 3.21 indicates that
majority of the lecturers disagree with this statement
implying that they believe that qualitative findings
are generalisable, which reflects a positivist mindset
on research findings. It is important to note that
some of the statements are on fundamental aspects
which differentiate a positivist research paradigm
from a interpretivist paradigm. The following section
provides some examples on this:
Statement 3 (see Table 3) is a typical underlying
concept for a positivist research paradigm. Researchers
who use logical positivism or quantitative research are
accustomed to using familiar terms such as reliability,
validity and objectivity in judging the quality of such
research. Thus, from such a point of view, qualitative
research may seem unscientific and anecdotal.
However, applying the same criteria used to judge
quantitative research on qualitative research, has
been extensively criticised as inappropriate (Healy
VOLUME Six NUMBER two july - december 2008 PLATFORM
171
OTHER AREAS
Table4: Statement 3 – A good qualitative research must be
reliable, valid and objective
Valid
0
5
6
7
Total
Frequency
4
2
12
15
33
Percent
12.1
6.1
36.4
45.5
100.0
Valid Percent
12.1
6.1
36.4
45.5
100.0
Cumulative
Percent
12.1
18.2
54.5
100.0
Table 5: Statement 5 – In qualitative research, the truth should be
established logically or supported by empirical evidence
Valid
Frequency
Percent
Valid Percent
0
2
3
4
5
6
7
Total
4
1
1
3
2
11
11
33
12.1
3.0
3.0
9.1
6.1
33.3
33.3
100.0
12.1
3.0
3.0
9.1
6.1
33.3
33.3
100.0
Cumulative
Percent
12.1
15.2
18.2
27.3
33.3
66.7
100.0
Table 6: Statement 8 – The quality of the research is determined
by the validity of the research
Valid
0
4
5
6
7
Total
Frequency
5
1
6
12
9
33
Percent
Valid Percent
15.2
3.0
18.2
36.4
27.3
100.0
15.2
3.0
18.2
36.4
27.3
100.0
Cumulative
Percent
15.2
18.2
36.4
72.7
100.0
and Perry, 2000; Lincon and Guba, 1985; Stenbacka,
2001). Table 4 below, shows that 88% of the lecturers’
responses ranged from 5 and 7, while the rest did not
attempt to rate this statement. Thus, the findings
here show that the lecturers in this study are more
accustomed to positivist research paradigm.
Statements 5 and 8 are other typical aspects underlying
research in the positivist research paradigm. For
Statement 5 (see Table 5), majority of the lecturers’
(72.7%) responses ranged from 5 to 7 which again
reflect a positivist research paradigm. Again for
Statement 8, as stated earlier the term “valid” typically
reflects a positivist research paradigm. For this
statement (see Table 6), the majority of the lecturers’
responses (81.9%) ranged from 5 to 7 on the Likert
scale. This again confirms that the lecturers definitely
come from a positivist research paradigm because
the non-positivist paradigm relies on personal views,
experiences and subjectivity in understanding human
actions and behaviour in research.
Research methods
There were 4 statements (see Table 7) in the
questionnaire which were used to gauge the lecturers’
perception regarding the methods used in qualitative
based research.
Statements 4, 25 and 26 were all concerned with
determining the number of participants needed in a
research to yield credible conclusions. Although it is
difficult to determine the correct sample size, for many
researchers using statistical analysis for their data, ‘a
sample size of thirty’ is held as the minimum number
of cases needed (Cohen, Manion and Morrison,
2000:93). Nevertheless, there is a guide to determine
the adequate number of cases for statistical analysis,
to ensure the rigour and robustness of the results.
However, this requirement is held differently in
qualitative based research. In some qualitative studies
on life history, one participant can be the sample
of the study. Furthermore, a case study approach
‘involves the investigation of a relatively small number
of naturally occurring (rather than researcher-created)
Table 7. Lecturers’ perception regarding research methods
No.
Statements
Mean Scores
4
A qualitative study can comprise of in-depth interviews with a single participant
3.15
15
The key element of research is to explain a phenomena by collecting data that can be
analysed using mathematically based methods (in particular statistics)
4.52
25
Doing a case study in qualitative research can involve a single case
3.3
26
The number of subjects interviewed in a research is important (must have adequate
number of participants) to come up with credible conclusions.
4.82
172
PLATFORM VOLUME Six NUMBER TWO july - december 2008
OTHER AREAS
Researcher’s Role
Table 8: Statement 15
Valid
0
3
4
5
6
7
Total
Frequency
4
3
3
12
9
2
33
Percent
Valid Percent
12.1
9.1
9.1
36.4
27.3
6.1
100.0
12.1
9.1
9.1
36.4
27.3
6.1
100.0
Cumulative
Percent
12.1
21.2
30.3
66.7
93.9
100.0
In a qualitative based research, the role of the researcher
is important. The ‘theoretical sensitivity’ (Strauss and
Corbin, 1990) of the researcher is pertinent in carrying
out a qualitative based research:
Theoretical sensitivity referred to the personal quality
of the researcher. It could indicate an awareness of
the subtleties of meaning of data
… [It] refers to the attribute of having insight, the
ability to give meaning to data, the capacity
to understand, and capability to separate the
pertinent from that which isn’t.
Strauss and Corbin, 1990:42
Table 9: Statement 26
Valid
0
2
3
4
5
6
7
Total
Frequency
5
2
1
2
3
11
9
33
Percent
Valid Percent
15.2
6.1
3.0
6.1
9.1
33.3
27.3
100.0
15.2
6.1
3.0
6.1
9.1
33.3
27.3
100.0
Cumulative
Percent
15.2
21.2
24.2
30.3
39.4
72.7
100.0
cases1’ (original emphasis, Hammersley, 1992:185).
From the responses of the lecturers, their mean scores
for Statements 4 and 25 were 3.15 and 3.3 respectively.
This indicated that the lecturers disagree that a single
participant or a single case is adequate for research
purposes. Again this is contradictory with the
interpretivist research paradigm and thus, confirms
that the lecturers in this study are certainly positivist.
This is also supported by the lecturers’ responses for
Statements 15 and 26 (refer to Table 6) where the
mean scores were 4.52 and 4.82 (indicating a more
agreeable nature) respectively. For both of these
statements, more than 69% (see Table 8 and Table 9)
of the lecturers’ responses ranged from 5 to 7 which
strongly implies that they believe in the positivist view
of research where credible research is determined by
the sample size.
Thus, in qualitative research, the researcher would be
subjectively immersed in the subject matter whereas
a quantitative researcher would tend to remain
objectively separated from the subject matter (Miles
and Huberman, 1994). However, the mean scores
(Table 10) indicated that the mindset of the lecturers in
this study was biased to a quantitative based research
which encouraged the researcher to be objective and
separated from the research.
Style of Presentation
The use of personal pronouns such as ‘I’ and ‘We’ are
forbidden in presenting research from the positivist
research paradigm. Scientific and technical writing
are usually objective and impersonal. However,
in qualitative research the presence of the author
is an important aspect of the research. Therefore,
subjectively based qualitative research usually is written
using personal pronouns, especially the pronoun ‘I’ to
highlight the researcher’s subjective rhetorical stance.
Table 10. Statement on Researcher’s Role
1
No.
Statements
Mean Scores
12
A researcher must be objective
5.39
17
The findings produced are influenced by the beliefs of the researcher.
3.36
21
A researcher’s own opinion must not interfere with the ‘subjects’ opinion.
6.3
Case studies may involve a single case.
VOLUME Six NUMBER two july - december 2008 PLATFORM
173
OTHER AREAS
Table 11. Lecturers’ presentation style for qualitative based research
No.
Statements
Mean Scores
9
A qualitative thesis can be told in a form of a story.
3.06
13
In writing a qualitative based research paper, using personal pronouns such as ‘I’ to
refer to the researcher is permitted.
3.03
Hyland (2003:256&257) states that using the pronoun
‘I’ emphasizes personal engagement of the author
with the audience, which is ‘an extremely valuable
strategy when probing connections between entities
that are generally more particular, less precisely
measurable, and less clear-cut’. Therefore, the use
of personal pronouns supports qualitative research
inquiry. Furthermore, ‘qualitative research excels at
‘telling the story’ from the participant’s viewpoint’
which provides a rich descriptive detail as compared
to quantitative based research (Trochim, 2006:3).
The lecturers in this study however are only familiar
with the styles that reflect a positivist research
paradigm which underpins the style of writing for
scientific and technical research documents. This
is clearly reflected in the mean scores (see Table 11)
where the lecturers disagree with the statements
which basically reflect a typical style of a qualitative
researcher.
Discussion
In this section, we revisit the research questions we
posed in this paper. We begin by discussing our
findings based on our first research question:
1. How familiar are the researchers from the
engineering discipline, in subjective, qualitative
based education research?
This study set out to investigate how familiar are the
lecturers from the engineering and technological
discipline in qualitative based research. It is widely
known that researchers from the scientific and
technological discipline are from the positivist
(quantitative) research paradigm. However, in this
study the lecturers were asked to evaluate statements
174
to gauge their familiarity with qualitative based
research which is often based on an interpretivist
(qualitative) research paradigm. The main objective is
to examine positivist lecturers’ perception regarding
qualitative based research that is mainly subjective in
nature. This is very important because researchers
may have different research paradigms, especially
those who are from diverse disciplines.
The findings in this paper clearly indicate that the
majority of researchers in this study reflect a positivist
research mindset. Thus, they are not familiar with
the fundamental knowledge underlying qualitative
based research which is subjective in nature. Thus,
research from a conflicting research paradigm such
as the interpretivist, would not be evaluated fairly
by these researchers. Furthermore, this study shows
that the lecturers in this study perceive a qualitative
research which is often subjective, based on their
understanding and familiarity of quantitative based
research. Obviously, a positivist researcher would not
agree with the underlying principles that guide an
interpretivist researcher.
As stated earlier, in the university where this research is
carried out, research papers from different disciplines
are evaluated by researchers who may have different
research paradigms. Thus, for example, a qualitative
research that highlights in depth qualitative interviews
of only one or two participants may not favour well
if the reviewer of the research is a positivist who is
well versed in quantitative approaches. The reviewer
would be very concerned with the number of subjects
used in the research. From a positivist point of view,
the reviewer’s concerns are justifiable. However, as
discussed in this paper, the quality and the validity
of a qualitative based research is not determined by
the number of participants in a research. Thus, the
PLATFORM VOLUME Six NUMBER TWO july - december 2008
OTHER AREAS
reviewer’s concern for the number of participants
used in a qualitative research would be regarded as
unnecessary. Furthermore, a positivist reviewer might
expect conclusive and tangible conclusions. What the
reviewer would not be aware is that the purpose of
a qualitative research is to ‘describe or understand
the phenomena of interest from the participants’
eyes’ (Guba and Lincon, 1999). Thus, it calls for a rich
description of the real life research context in which the
complexities of different individuals who participated
in the research, is acknowledged and explored.
According to Creswell (1994:6), it is the responsibility of
the researcher to ‘report faithfully these realities and to
rely on voices and interpretations of informants’. Thus
there is ‘no true or valid interpretation’, only ‘useful,
liberating, fulfilling and rewarding’ interpretations
(Crotty, 1998:48).
Conclusion
Therefore, the findings in this study supports the
discussion above, that it is not appropriate for a
researcher from a conflicting research paradigm to
evaluate research papers from researchers who do not
share the same research paradigm as the evaluators.
This study clearly supports that researchers with
different research paradigms view research differently.
It is inappropriate to impose one’s underlying
principles of research on others, especially if the
research reflects contradictory research paradigms
such as the positivist and interpretivist as shown in
this study.
• Research methods
• Researcher’s role
• Style of presentation
Although this paper started with two research
questions, we did not deliberate on the second
research question:
2. What are the engineering researchers’ perceptions
regarding the contribution of qualitative based
research?
This is because the study showed that the lecturers
are not familiar with qualitative based research and
thus will not be able to contribute to this question.
This paper highlights that different research paradigms
reflect differences in knowledge on the underlying
principles that guide researchers. This study confirms
that the researchers in this study have a positivist
research mindset. However, this study also showed
how researchers from the positivist mindset radically
differ from researchers from the non-positivist
(interpretivist) research paradigm.
Furthermore,
the findings in this study indicate that the lecturers
impose their positivist knowledge even on research
from non-positivist (interpretivist) research paradigm.
Thus, this study stresses the importance of knowing
that researchers may differ in their research paradigms
and as a consequence, may also differ in the following
aspects regarding research:
Thus, even within a discipline, identifying the
researcher’s research paradigm is important in
determining the true value of a research.
References
[1]
Burns. (2000). “Introduction to research methods”. London:
Sage Publications
[2]
Cohen, L., Manion, L., & Morrison, K. (2000). “Research
Methods in Education” (5 ed.). London: Routledge Falmer
[3]
Creswell, J. W. (2003). “Research Design Qualitative,
Quantitative and Mixed Methods approaches”. London:
Sage Publication
[4]
Crotty, M. (1998). “The Foundations of Social Research Meaning and perspective in the research process”. London:
Sage Publications Ltd
[5]
Frankel, R. (2005). The “White Space” of Logistics Research:
A look at the role of methods usage. Journal of Business
Logistics, 1-12
[6]
Guba, E., & Lincon, Y. S. (1994). “Competing paradigms in
Qualitative Research”. In N. K. Denzin & Y. S. Lincon (Eds.),
Handbook of Qualitative Research (pp. 105-117). California:
Sage Publication
[7]
Hammersley, M. (1992). “What’s wrong with ethnography”.
London: Routledge
VOLUME Six NUMBER two july - december 2008 PLATFORM
175
OTHER AREAS
[8]
Healy, M., & Perry, C. (2000). “Comprehensive criteria to
judge validity and reliability of qualitative research within
the realism papradigm”. Qualitative Market Research, 3(3),
118-126
[9]
Hitchcock, G., & Hughes, D. (1995). “Research and the
teacher”. London: Routledge
[24] Strauss, A., & Corbin, J. (1990). “Basics of qualitative research:
Grounded theory procedures and techniques”. Newbury
Park: Sage Publications
[25] Trochim, M. K. (2006, 20 October 2006). “Qualitative Validity”.
Retrieved 20 January, 2007, from the World Wide Web:
http://www.socialresearchmethods.net/kb/qualval.php
[10] Hollis, M. (1994). “The Philosophy of Social Science: An
Introduction”. Cambridge: Cambridge University Press
[11] Howe, K. R. (2003). “Closing methodological divides: Toward
democratic educational research” (Vol. 100). Dordrecht,
Netherlands: Kluwer
Sumathi Renganathan is a senior lecturer
in the Management and Humanities
Department. Her areas of specialisation
and interest are: Language in Education,
Second Language Socialisation, Multilingualism, and Language and Ethnicity.
[12] Hussey, J., & Hussey, R. (1997). “Business Research”. London:
Mac Millan Press Ltd
[13] Hyland, K. (2003). “Self-Citation and Self-Reference:
Credibility and Promotion in Academic Publication”. Journal
of the American Society for Information Science and
Technology, 54(3), 25 -259
[14] Johl, S. K., Bruce, A., & Binks, M. (2006). “Using mixed
method approach in conducting a business research”.
Paper presented at the 2nd International Borneo Business
Conference, Kuching, Sarawak
[15] King, G., Keohane, R., & Verba, S. (1994). “Designing Social
Inquiry: Scientific Inferences in Qualitative Research”. New
Jersey: Princeton University Press
[16] Lincon, Y. S., & Guba, E. G. (1985). “Naturalistic inquiry”.
Beverly Hills, CA: Sage Publications
[17] Love, P. E. D., Holt, G. D., & Heng, L. (2002). “Triangulation
in Construction Management Research”. Engineering
Construction and Architecture Management, 9(4), 294-303
[18] McGrath, J. E. (1982). “Dilemmatics: The Study of Research
Choices and Dilemmas”. In J. E. McGrath & J. Martin & R.
KuIa (Eds.), Judgment Calls in Research. Beverly Hills: Sage
Publications
[19] Miles, M., & Huberman, M. A. (1994). “Qualitative data
analysis”. Beverly Hills: Sage Publications
[20] Moser, P., & Trout, J. D. (Eds.). (1995). “Contemporary
materialism”. London: Routledge
[21] Neuman, W. L. (2000). “Social Research Methods Qualitative
and Quantitative Approaches”. Boston: Pearson Education
Company
[22] Sekaran, U. (2000). “Research methods for business, a skillbuilding approach”. New York: John Wiley and Sons Inc
[23] Stenbacka, C. (2001). “Qualitative research requires quality
concepts of its own”. Management Decision, 39(7), 551-555
176
PLATFORM VOLUME Six NUMBER TWO july - december 2008
Satirenjit Kaur is a senior lecturer in the
Management and Humanities Department.
Her areas of specialisation and interests are
entrepreneurship and management.
PL ATFORM is a biannual, peerreviewed journal of Universiti Teknologi
PETRONAS. It serves as a medium for
faculty members, students and industry
professionals to share their knowledge,
views, experiences and discoveries in
their areas of interest and expertise. It
comprises collections of, but not limited
to, papers presented by the academic
staff of the University at various local and
international conferences, conventions
and seminars.
The entries range from opinions and
views on engineering, technology and
social issues to deliberations on the
progress and outcomes of academic
research.
Opinions expressed in this journal need
not necessarily reflect the official views
of the University.
All materials are copyright of Universiti
Teknologi PETRONAS. Reproduction
in whole or in part is not permitted
without the written permission from
the University.
Notes for Contributors
Instructions to Authors
Authors of articles that fit the aims, scopes and policies of
this journal are invited to submit soft and hard copies to
the editor. Paper should be written in English. Authors are
encouraged to obtain assistance in the writing and editing
of their papers prior to submission. For papers presented or
published elsewhere, authors should include the details of
the conference or seminar.
Manuscript should be prepared in accordance with the
following:
1.The text should be preceded by a short abstract of 50-100
words and four or so keywords.
2.The manuscript must be typed on one side of the
paper, double-spaced throughout with wide margins
not exceeding 3,500 words although exceptions will be
made.
3. Figures and tables have to be labelled and should be
included in the text. Authors are advised to refer to
recent issues of the journals to obtain the format for
references.
4. Footnotes should be kept to a minimum and be as brief
as possible; they must be numbered consecutively.
5. Special care should be given to the preparation of the
drawings for the figures and diagrams. Except for a
reduction in size, they will appear in the final printing in
exactly the same form as submitted by the author.
6. Reference should be indicated by the authors’ last names
and year of publications.
PLATFORM Editor-in-Chief
Universiti Teknologi PETRONAS
Bandar Seri Iskandar
31750 Tronoh
Perak Darul Ridzuan
MALAYSIA
VOLUME SIX NUMB ER TWO J UL Y - D ECEMBER 2008
P L AT F O R M
P L AT F O R M
Volume 6 Number 2
2
6
13
21
27
31
38
47
52
65
77
85
91
96
105
111
116
122
129
137
145
152
158
166
VOLUME SIX NU MB ER TWO JU L Y - D EC EMB ER 2008
Mission-Oriented Research: CARBON DIOXIDE MANAGEMENT
Separation Of Nitrogen From Natural Gas By Nano- Porous Membrane Using Capillary Condensation
Farooq Ahmad, Hilmi Mukhtar, Zakaria Man, Binay. K. Dutta
Mission-Oriented Research: DEEPWATER TECHNOLOGY
Recent Developments In Autonomous Underwater Vehicle (AUV) Control Systems
Kamarudin Shehabuddeen, Fakhruldin Mohd Hashim
Mission-Oriented Research: GREEN TECHNOLOGY
Enhancement Of Heat Transfer Of A Liquid Refrigerant In Transition Flow In The Annulus Of A DoubleTube Condenser R. Tiruselvam, Chin Wai Meng, Vijay R Raghavan
Mission-Oriented Research: PETROCHEMICAL CATALYSIS TECHNOLOGY
Fenton And Photo-Fenton Oxidation Of Diisopropanolamine
Abdul Aziz Omar, Putri Nadzrul Faizura Megat Khamaruddin, Raihan Mahirah Ramli
Synthesis Of Well-Defined Iron Nanoparticles On A Spherical Model Support
Noor Asmawati Mohd Zabidi, P. Moodley, P. C. Thüne, J. W. Niemantsverdriet
Technology Platform: FUEL COMBUSTION
Performance And Emission Comparison Of A Direct-Injection (DI) Internal Combustion Engine Using
Hydrogen And Compressed Natural Gas As Fuels
Abdul Rashid Abdul Aziz, M. Adlan A., M. Faisal A. Mutalib
The Effect Of Droplets On Buoyancy In Very Rich Iso-Octane-Air Flames
Shaharin Anwar Sulaiman, Malcolm Lawes
Technology Platform: SYSTEM OPTIMISATION
Anaerobic Co-Digestion Of Kitchen Waste And Sewage Sludge For Producing Biogas
Amirhossein Malakahmad, Noor Ezlin Ahmad Basri, Sharom Md Zain
On-Line At-Risk Behaviour Analysis And Improvement System (E-ARBAIS)
Azmi Mohd Shariff, Tan Sew Keng
Bayesian Inversion Of Proof Pile Test: Monte Carlo Simulation Approach
Indra Sati Hamonangan Harahap, Wong Chun Wah
Element Optimisation Techniques In Multiple DB Bridge Projects
Narayanan Sambu Potty, C. T. Ramanathan
A Simulation Study On Dynamics And Control Of A Refrigerated Gas Plant
Nooryusmiza Yusoff, M. Ramasamy, Suzana Yusup
Technology Platform: APPLICATION OF INTELLIGENT IT SYSTEM
An Interactive Approach To Curve Framing Abas Md Said
Student Industrial Internship Web Portal Aliza Sarlan, Wan Fatimah Wan Ahmad, Dismas Bismo
Hand Gesture Recognition: Sign To Voice System (S2V) Foong Oi Mean, Tan Jung Low, Satrio Wibowo
Parallelization Of Prime Number Generation Using Message Passing Interface
Izzatdin A Aziz, Nazleeni Haron, Low Tan Jung, Wan Rahaya Wan Dagang
Evaluation Of Lossless Image Compression For Ultrasound Images
Boshara M. Arshin, P. A. Venkatachalam, Ahmad Fadzil Mohd Hani
Learning Style Inventory System: A Study To Improve Learning Programming Subject
Saipudnizam Mahamad, Syarifah Bahiyah Rahayu Syed Mansor, Hasiah Mohamed
Performance Measurement – A Balanced Score Card Approach
P. D. D. Dominic, M. Punniyamoorthy, Savita K Sugathan, Noreen I. A.
A Conceptual Framework For Teaching Technical Writing Using 3D Virtual Reality Technology
Shahrina Md Nordin, Suziah Sulaiman, Dayang Rohaya Awang Rambli,
Wan Fatimah Wan Ahmad, Ahmad Kamil Mahmood
Multi-Scale Color Image Enhancement Using Contourlet Transform
Melkamu H. Asmare, Vijanth Sagayan Asirvadam, Lila Iznita
Automated Personality Inventory System Wan Fatimah Wan Ahmad, Aliza Sarlan, Mohd Azizie Sidek
A Fuzzy Neural Based Data Classification System Yong Suet Peng, Luong Trung Tuan
Other Areas
Research In Education: Taking Subjective Based Research Seriously
Sumathi Renganathan, Satirenjit Kaur
Jul - Dec 2008
I SSN 1 5 1 1 - 6 7 9 4