p - CITEE 2015 - Universitas Gadjah Mada

Transcription

p - CITEE 2015 - Universitas Gadjah Mada
Number 2
ISSN: 2085-6350
PROCEEDINGS OF
CONFERENCE ON
INFORMATION TECHNOLOGY
AND ELECTRICAL ENGINEERING
INTERNATIONAL SESSION
Signals, Systems, and Circuits
DEPARTMENT OF ELECTRICAL ENGINEERING
FACULTY OF ENGINEERING
GADJAH MADA UNIVERSITY
Conference on Information Technology and Electrical Engineering (CITEE)
Organizer
Steering Commitee
• Adhi Susanto (UGM)
• Hamzah Berahim (UGM)
• Thomas Sri Widodo (UGM)
• Dadang Gunawan (UI)
• Heri Mauridi (ITS)
• Yanuarsyah Harun (ITB)
• Anto Satrio Nugroho (BPPT)
• Son Kuswadi (PENS)
Advisory Board
• Tumiran (UGM)
• Lukito Edi Nugroho (UGM)
• Anto Satrio Nugroho (BPPT)
• Son Kuswadi (PENS)
General Chair
• Bambang Sutopo
Organizing Chairs
• Risanuri Hidayat
• Sri Suning Kusumawardhani
• Ridi Ferdiana
• Adha Imam Cahyadi
• Budi Setiyanto
Program Chairs
• Prapto Nugroho
• Agus Bejo
• Cuk Supriyadi Ali Nandar (BPPT)
• Yusuf Susilo Wijoyo
Publication Chair
• Enas Dhuhri K
Finance Chairs
• Eny Sukani Rahayu
• Maun Budiyanto
• Roni Irnawan
Secretariats
• Astria Nur Irfansyah
• Lilik Suyanti
YOGYAKARTA, AUGUST 4, 2009
Conference on Information Technology and Electrical Engineering (CITEE) 2009
FOREWORD
First of all, praise to Almighty God, for blessing us with healthy and ability to come here, in the
Conference of Information and Electrical Engineering 2009 (CITEE 2009). If there is some noticeable wisdoms
and knowledge must come from Him.
I would like to say thank you to all of the writers, who come here enthusiastically to share experiences and
knowledge. Without your contribution, this conference will not has a meaning.
I also would like to say thank you to Prof. Dadang Gunawan from Electrical Engineering, University of
Indonesia (UI), Prof. Yanuarsyah Haroen from Electrical Engineering and Informatics School, Bandung Institute
of Technology, ITB, Prof. Mauridhi Hery Purnomo from Electrical Engineering Department, Surabaya Institute
of Technology (ITS). And also Prof. Takashi Hiyama from Kamamoto University, Japan, Thank you for your
participation and contribution as keynote speakers in this conference.
This conference is the first annual conference held by Electrical Engineering Department, Gadjah Mada
University. We hope, in the future, it becomes a conference of academics and industries researchers in the field of
Information Technology and Electrical Engineering around the world. We confine that if we can combine these
two fields of sciences, it would make a greater impact on human life quality.
According to our data, there are 140 writers gather here to present their papers. They will present 122 titles
of papers. There are 47 papers in the field of Electrical Power Systems, 53 papers in the area of Systems, Signals
and Circuits, and 22 papers in Information Technology. Most of these papers are from universities researchers.
We hope, the result of the proceedings of this conference can be used as reference for the academic and
practitioner researchers to gain
At last, I would like to say thank you to all of the committee members, who worked hard to prepare this
conference. Special thanks to Electrical Engineering Department, Gadjah Mada University, of supporting on
facilities and funds.
Thank you and enjoy the conference, CITEE 2009, and the city, Yogyakarta
August, 4Th, 2009
Bambang Sutopo
Electrical Engineering Dept., Fac. of Engineering, GMU
Proceedings of CITEE 2009
Number 2
ISSN: 2085-6350
Table of Contents
Organizer
Foreword
Table of Contents
Schedule
ii
iii
v
vii
KEYNOTE
Social Intelligent on Humanoid Robot: Understanding Indonesian Text Case Study
Mauridhi Hery Purnomo (Electrical Engineering Department, ITS, Indonesia)
1
Signal Processing: Video Compression Techniques
Dadang Gunawan (Electrical Engineering Department, University of Indonesia)
4
Intelligent Systems Application to Power Systems
Prof. Takashi Hiyama (Kumamoto University, Japan)
-
INTERNATIONAL SESSION: Signals, Systems, and Circuits
Analysis and Simulation of Bus-Clamping PWM Techniques Based on Space Vector Approach
Ms.M.Thanuja, Mrs. K. SreeGowri
7
Medical Image Processing of Proximal Femur X-Ray for Osteoporosis Detection
Riandini, Mera Kartika Delimayanti, Donny Danudirdjo
15
Performance Evaluation of Free-Space Optical Communication System on Microcell Networks in Urban Area
Purnomo Sidi Priambodo, Harry Sudibyo and Gunawan Wibisono
22
A Measure of Vulnerability for Communication Networks: Component Order Edge Connectivity
A. Suhartomo, Ph.D.
27
The Speech Coder at 4 kbps using Segment between Consecutive Peaks based on the Sinusoidal Model
Florentinus Budi Setiawan
31
Automatic Abnormal Waves Detection from the Electroencephalograms of Petit Mal Epilepsy Cases to Sort Out the Spikes Fp1 Fp2, the Sharps, the Polyphase Based on Their Statistical Zerocrossing
Siswandari N, Adhi Susanto, Zainal Muttaqin
35
Studies on the Limitation of Fighter Aircraft Maneuver Caused by Automatic Control Design
Okky Freeza Prana Ghita Daulay, Arwin Datumaya Wahyudi Sumari
39
Feature Extraction and Selection on Osteoporosis X-Ray Image for Content Based Image Retrieval (CBIR) Purposes
Usman Balugu, Ratnasari Nur Rohmah, Nurokhim
45
The Implementation of Turbo Encoder and Decoder Based on FPGA
Sri Suning Kusumawardani and Bambang Sutopo
51
BER Performance Analysis of PAM and PSM for UWB Communication
Risanuri Hidayat
55
Hardware Model Implementation of a Configurable QAM Mapper-Demapper for an Adaptive Modulation OFDM System
Budi Setiyanto, Astria Nur Irfansyah, and Risanuri Hidayat
61
Hardware Model Implementation of a Baseband Conversion, Chip Synchronization, and Carrier Synchronization Technique for a
Universal QAM System
Budi Setiyanto, Mulyana, and Risanuri Hidayat
69
Comparison Study of Breast Thermography and Breast Ultrasonography for Detection of Breast Cancer
Thomas Sri Widodo, Maesadji Tjokronegore, D. Jekke Mamahit
75
Adaptive Polynomial Approximation for Gravimetric Geoid: A case study using EGM96 and EIGEN-GL04C Geopotential
Development
Tarsisius Aris Sunantyo, Muhamad Iradat Achmad
78
Cerebellar Model Associative Computer (CMAC) for Gravimetric Geoid study based on EGM96 and EIGEN-GL04C Geopotential
Development
Muhamad Iradat Achmad, Tarsisius Aris Sunantyo, Adhi Susanto
83
Conference on Information Technology and Electrical Engineering (CITEE)
Conference on Information Technology and Electrical Engineering (CITEE)
SCHEDULE
Tuesday, August 4, 2009
07.30 – 08.00:
08.00 – 08.15:
Registration
Opening
1. Welcome speech by conference chairman
2. Speech by GMU’s Rector
08.15 – 09.20:
PLENARY SESSION
Prof. Takashi Hiyama (Kumamoto University, Japan): Intelligent Systems Application to
Power Systems
Prof. Dr. Mauridhi Hery Purnomo (Electrical Engineering Department,ITS, Indonesia):
Social Intelligent on Humanoid Robot: Understanding Indonesian Text Case Study
Prof. Dr. Dadang Gunawan (Electrical Engineering Department, University of
Indonesia): Signal Processing: Video Compression Techniques
Prof. Dr. Yanuarsyah Haroen (Electrical Engineering and Informatics School, ITB,
Indonesia): Teknologi Sistem Penggerak dalam WahanaTransportasi Elektrik
09.20 – 09.30: Break
PARALLEL SESSION
INTERNATIONAL SESSION (Room 1, 2)
Room: 1
Time
Group
Country/City
09.30 – 09.45
P
Malaysia
09.45 – 10.00
10.00 – 10.15
10.15 – 10.30
10.30 – 10.45
10.45 – 11.00
11.00 – 11.15
11.15 – 11.30
11.30 – 11.45
11.45 – 12.00
12.00 – 13.00
13.00 – 13.45
13.15 – 13.30
P
P
P
Papua
Medan
Bandung
P
P
P
P
P
Bandung
Surabaya
Surabaya
Surabaya
Yogyakarta
P
P
Surabaya
Surabaya
13.30 – 13.45
I
Surabaya
13.45 – 14.00
14.00 – 14.15
14.15 – 14.30
14.30 – 14.45
14.45 – 15.00
15.00 – 15.15
15.15 – 15.30
I
I
Surabaya
Surabaya
I
I
P
P
Yogyakarta
Surakarta
Yogyakarta
Yogyakarta
Author(s) or Presenter(s)
Zulkarnain Lubis, Ahmed N. Abdalla, Samsi bin MD said .Mortaza bin
Mohamed
Adelhard Beni Rehiara
Zakarias Situmorang
Kartono Wijayanto, Yanuarsyah Haroen
Coffee Break
Hermagasantos Zein
Ali Musyafa, Soedibjo, I Made Yulistiya Negara , Imam Robandi
Buyung Baskoro, Adi Soeprijanto, Ontoseno Penangsang
Eko Prasetyo, Boy Sandra, Adi Soeprijanto
T. Haryono, Sirait K.T., Tumiran, Hamzah Berahim
Lunch Break
Dimas Anton Asfani, Nalendra Permana
Dimas Anton Asfani, Iman Kurniawan, Adi Soeprijanto
F.X. Ferdinandus, Gunawan, Tri Kurniawan Wijaya, Novita Angelina
Sugianto
Arya Tandy Hermawan, Gunawan, Tri Kurniawan Wijaya
Herman Budianto, Gunawan, Tri Kurniawan Wijaya, Eva Paulina Tjendra
Coffee Break
Bambang Soelistijanto
Munifah, Lukito Edi Nugroho, Paulus Insap Santosa
Nurcahyanto, T. Haryono, Suharyanto.
Agni Sinatria Putra, Tiyono, Astria Nur Irfansyah
Notes:
1. P: Electrical Power Systems; S: Signals, Systems, and Circuits; I: Information Technology
2. Paper titles are listed in Table of Contents
Department of Electrical Engineering, Faculty of Engineering, Gadjah Mada University
Conference on Information Technology and Electrical Engineering (CITEE)
Room: 2
Time
09.30 – 09.45
09.45 – 10.00
10.00 – 10.15
10.15 – 10.30
10.30 – 10.45
10.45 – 11.00
11.00 – 11.15
11.15 – 11.30
11.30 – 11.45
11.45 – 12.00
12.00 – 13.00
13.00 – 13.45
13.15 – 13.30
13.30 – 13.45
13.45 – 14.00
14.00 – 14.15
14.15 – 14.30
14.30 – 14.45
14.45 – 15.00
15.00 – 15.15
15.15 – 15.30
Group
S
S
S
S
Country/City
INDIA
Jakarta
Jakarta
Jakarta
I
I
I
S
S
Yogyakarta
Yogyakarta
Lampung
Semarang
Semarang
S
S
S
S
S
Yogyakarta
Yogyakarta
Yogyakarta
Yogyakarta
Yogyakarta
S
S
S
S
Yogyakarta
Yogyakarta
Yogyakarta
Yogyakarta
Author(s) or Presenter(s)
Ms.M.Thanuja, Mrs. K. SreeGowri
A. Suhartomo
Riandini, Mera Kartika Delimayanti, Donny Danudirdjo
Purnomo Sidi Priambodo, Harry Sudibyo and Gunawan Wibisono
Coffee Break
Arwin Datumaya Wahyudi Sumari, Adang Suwandi Ahmad
Arwin Datumaya Wahyudi Sumari, Adang Suwandi Ahmad
Sumadi, S; Kurniawan, E.
Florentinus Budi Setiawan
Siswandari N, Adhi Susanto, Zainal Muttaqin
Lunch Break
Thomas Sri Widodo, Maesadji Tjokronegore, D. Jekke Mamahit
Tarsisius Aris Sunantyo, Muhamad Iradat Achmad
Muhamad Iradat Achmad, Tarsisius Aris Sunantyo, Adhi Susanto
Usman Balugu, Ratnasari Nur Rohmah, Nurokhim
Okky Freeza Prana Ghita Daulay, Arwin Datumaya Wahyudi Sumari
Coffee Break
Sri Suning Kusumawardani and Bambang Sutopo
Risanuri Hidayat
Budi Setiyanto, Astria Nur Irfansyah, and Risanuri Hidayat
Budi Setiyanto, Mulyana, and Risanuri Hidayat
NATIONAL SESSION (Room 3, 4, 5, 6, 7)
Yogyakarta, August 4, 2009
Proceedings of CITEE, August 4, 2009
-1
Keynote
Social Intelligent on Humanoid Robot:
Understanding Indonesian Text
Case Study
Mauridhi Hery Purnomo
Electrical Engineering Department-Institut Teknologi Sepuluh November
Surabaya 60111, Indonesia
[email protected]
Abstract— Social affective and emotion are required on
humanoid robot performance to make the robot be more
human. Social intelligent are the individual ability to manage
relationship with other agents and act wisely based on
previous learning experiences. Here, social intelligent is
intended to understand Indonesian text. How the computation
process, as well as affective interaction, emotion expression of
the humanoid robot to the human statement. This process is a
highly adaptive complex approximation, dependently on its
entire situation and environment.
Keywords—social, affective, emotion, intelligent, computing
(key words)
I. INTRODUCTION
Social and interactive behaviors are necessary
requirements in wide implementation areas and contexts
where robots need to interact and collaborate with other
robots or humans. The nature of interactivity and social
behavior in robot and humans is a complex model.
An experimental robot platform KOBIE, which provides
a simulation tool for emotion expression system includes an
emotion engine was developed. The simulation tool
provides a visualization interface for the emotion engine
and expresses emotion through an avatar. The system can be
used in the development of cyber characters that use
emotions or in the development of an apparatus with
emotion in a ubiquitous environment [1].
To improve the understandability and friendliness in
human-computer interfaces and media contents, a
Multimodal Presentation Markup Language (MPML) is
developed. MPML is a simple script language to make
multi-modal presentation contents using animated
characters for presenters [2].
Other effort in the robot head which uses arm-type
antennae, eye-expression, and additional exaggerating parts
for dynamic emotional expression is also developed. The
robot head is developed for various and efficient emotional
expressions in the Human-Robot interaction field. The
concept design of the robot is an insect character [3].
In regard to artificial cognitive, iCub humanoid robot
systems is developed. The system is open-systems 53
degree-of-freedom cognitive humanoid robot, 94 cm tall,
the same size as a three year-old child. Able to crawl on all
fours and sit up, its hands will allow dexterous
manipulation, and its head and eyes are fully articulated. It
has visual, vestibular, auditory, and haptic sensory
capabilities [4].
An innovative integration of interactive group learning,
multimedia technology, and creativity used to enhance the
learning of basic psychological principles was created.
This system is based on current robotic ideology calling for
the creation of a PowerPoint robot of the humanoid type
that embodies the basic theories and concepts contained in a
standard psychological description of a human being [5].
Now days, not only visual and auditory information are
used in media and interface fields but also multi-modal
contents including documents such like texts. Thus, in this
paper, a part of result on emotion expression and
environment through understanding Indonesian text, as
affective interaction between a human and a robot is
explored.
The paper is organized as follows. In Section 2, the
general emotion and expression system on life-like agent is
presented. Section 3 describes the experimental on
Indonesian text classification. In Section 4, the preliminary
result in emotion classification of Indonesian article text is
discussed.
II.
EMOTION AND EXPRESSION ON LIFE-LIKE AGENT
Social Computing, Social Agent, and Life-likeness
Many Psychologists have studied a definition and
classification of emotions, therefore, so many classification
methods of emotions and expressions. However, we need to
choose categories of emotions that are suitable expressed by
robot, as well as the well-known Ekman’s 6 basic emotion
expressions model can be used.
In social computing life-like characters are the key, and
the affective functions create believability. To articulate
synthetic emotions can be presented as; personalities, human
interactive behavior or presentation skills. The personalities;
by means of body movement, facial display, and the
coordination of the embodied conversational behavior of
multiple characters possibly including the user. Personality
is key to achieving life-likeness
Some Applications of Life-Like Character
Life-like characters are synthetic agents apparently living
on the screen of computers. Life-like character can be
implemented as virtual tutors and trainers in interactive
learning environments. On the web as an information expert,
presenter, communication partners, and enhancing the search
engine. The other application as actors for entertainment, in
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
2
Keynote
online communities and guidance systems as personal
representatives.
Early characterization of the emotional and believable
character was raised by Joseph Bates. He said, the portrayal
of emotions plays a key role in the aim to create believable
characters, one that provides the illusion of life, and thus
permits the audience’s suspension of disbelief. In game and
animation, suspension of disbelief is very important, for
instance as: synthetic actors, non-player characters, and
embodied conversational agents.
Proceedings of CITEE, August 4, 2009
Figure 1 is an example how to train machine (agent) in
order responsive to the external and adequate response. The
case study is, Indonesian text classification. There are two
types of machine learning, supervised and unsupervised
learning.
Figure 2 show a block diagram process of embodied
conversational agent.
IV.
Emotion and personality are often seen as the affective
bases of believability, and sometimes the broader term social
is used to characterize life-likeness.
•
III.
CLASSIFICATION SYSTEM FOR INDONESIAN TEXT
Information growth, including texts are faster than
human ability, thus help system is quite necessary. For
instance as the following illustration;
y
The Recent study, which used Web searches in 75
different languages to sample the Web, determined
that there were over 11.5 billion (1012) Web pages
in the publicly indexable Web as of the end of
January 2005
y
As of March 2009, the indexable web contains at
least 25.21 billion pages
y
On July 25, 2008, Google software engineers Jesse
Alpert and Nissan Hajaj announced that Google
Search had discovered one trillion unique URLs
y
As of May 2009, over 109.5 million websites
operated
feature
extractor
..features..
test
input
language
dependent
NLP tools
feature
extractor
..features..
machine
learning
classifier
model
•
“When I nearly walked on a blindworm and then
saw it crawl away” → disgust
•
“When I was involved in a traffic accident” → fear
•
“I do not help out enough at home” → guilt
•
“Passing an exam I did not expect to pass” → joy
•
“Failing an examination” → sadness
•
“When, as an adult I have been caught lying or
behaving badly” → shame
The classification is divided into six (6) classes of
emotion: disgust, shame, anger, sadness, joy and fear.
Each class has 200 text files, data: “as-is”; DataNot:
pre-processing only handles “not”. Split ratio 0.5 shows
f-measure scores 0.59
label
language
dependent
NLP tools
The following sentences are sample of statements,
and some emotion expressions;
“When a car is overtaking another and I am forced
to drive off the road” → anger
Based on some statements and the emotion expression as
mentioned above in Indonesian text, the preliminary
classification results are shown in the table 1, figure 3 and
figure 4.
(a) training phase
training
input
EMOTION CLASSIFICATION FROM INDONESIAN
ARTICLE TEXT
Pre-processing Steps
predicted
label
(b) prediction phase
Figure 1. Example of Indonesian Text Classification
Text Classification (TC) techniques usually ignore stopwords and case of input text. In pre-processing step, stopwords removal can be applied.
Stop-words such as “not”, “in”, “which” and exclamation
marks (“!”) usually do not affect categorization of text.
Text
Input
Knowledge
Base
Response
User
(Human)
Text-based
Conversational Agent
Information
Retrieval
Text
Classification
Text
Mining
Figure 2. Knowledge from Free (Unstructured) Text
The illustrations as mentioned above explain the
essential of help system especially in Indonesian text. We
have developed a system for understanding and classifying
Indonesian text, and the block diagram as shown in figure 1
and figure 2.
ISSN: 2085-6350
TABLE I. NAÏVE BAYES CLASSIFICATION INTO 4 CLASS
Free
Text
Classifica
tion ratio
(%)
Usual & original text
Text without stop
words
20
71.41
69.81
40
73.33
71.3
60
74.40
71.3
80
75.33
74.05
Accuracy (%)
Example:
-
Microsoft released Windows → categorized as
“news”.
-
Microsoft has not released Windows yet → still
categorized as “news”.
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
Input text is converted
classifications, as example:
to
-3
Keynote
lowercase
on
emotion
DataNot
-
“I do like you.” ≠ “I do not like you!”
70
-
“I do not like you.” ≠ “I DO NOT LIKE YOU!!”
65
60
F-Measure
Naive Bayes
70
55
50
65
Multinomial NB
45
NB
Precision
60
40
55
0
0,1
0,2
0,4
0,5
0,6
0,7
0,8
0,9
1
Rasio data
50
Data
DataNot
With pre processing
45
Figure 4. Results Recapitulation of Indonesian Text Classification
40
40
45
50
55
60
65
70
Recall
V.
Multinomial NB
70
65
60
55
REFERENCES
Data
DataNot
50
[1]
45
40
[2]
40
45
50
55
CONCLUDING REMARKS
The preliminary results of emotion expression and related
environment through Indonesian text are described. We
develop an Indonesian conversational agent system includes
an emotion expression engine, that will used in the game
engine. The use of emotion on Indonesian text is expected
to the improvement of expressiveness of understanding and
actions. The research are still underway, so many possibility
to improve our future works and making the system more
life-like.
non bayesian
Precision
0,3
60
65
70
Recall
[3]
Multinomial non bayesian
Figure 3. Emotion Classification of Indonesian Text
[4]
Data
65
[5]
60
C Park, J W Ryu, J Kim, S Kang, J Sohn, YJ Cho, “Emotion
Expression and Environment Through Affective Interaction”
Proceedings of the 17th World Congress The International Federation
of Automatic Control,Seoul, Korea, July 6-11, 2008 .
K Kushida, Y Nishimura, et al.“Humanoid Robot Presentation
through Multimodal Presentation Markup Language MPML-HR”
AAMAS’05, Utrecht, Netherlands, July 25-29, 2005.
H Song and DS Kwon, “Design of a Robot Head with Arm-type
Antennae for Emotional Expression”, International Conference on
Control, Automation and Systems in COEX, Seoul, Korea Oct. 1720, 2007.
G Sandini, G Metta, and D Vernon,”The iCub Cognitive Humanoid
Robot:An Open-System Research Platform for Enactive Cognition”,
M. Lungarella et al. (Eds.): 50 Years of AI, Festschrift, LNAI 4850,
Springer-Verlag Berlin Heidelberg, pp. 359–370, 2007.
James L. Anderson and Erin M. Applegarth, “The Psychological
Robot: A New Tool for Learning, 3rd ed., International Journal of
Teaching and Learning in Higher Education 2007, Volume 19,
Number 3, 305-314
F-Measure
55
50
Multinomial NB
45
Naive Bayes
40
0
0,1
0,2
0,3
0,4
0,5
0,6
0,7
0,8
0,9
1
Rasio data
Without pre processing
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
4
Keynote
Proceedings of CITEE, August 4, 2009
Signal Processing:
Video Compression Techniques
Dadang Gunawan
Electrical Engineering Department, University of Indonesia
In our information society, signal processing has been
created a significant effect. Signal processing can be found
everywhere: in home appliances, in Cell Phone, TVs,
Automobile, GPSs, Modem Scanner, and All kind of
Communication Systems and Electronic Devices. Modern
cell phones are indeed a most typical example – within
these small wonders, voice, audio, image, video and
graphics are processed and enhanced based on decades of
media signal processing research.
Technological advancement in recent years has
proclaimed a new golden age for signal processing [1].
Many exciting directions, such as bioinformatics, human
language, networking, and security, are emerging from
traditional field of signal processing on raw information
content. The challenge in the new era is to transcend from
the conventional role of processing in low level, waveformlike signal to the new role of understanding and mining the
high-level, human-centric semantic signal and information.
Such a fundamental shift has already taken place in limited
areas of signal processing and is expected to become more
pervasive in coming years of research in more areas of
signal processing.
One of the huge applications of signal processing is
exploited as video compression. Nowadays, video
applications such as digital laser disc, electronic camera,
videophone and video conferencing systems, image and
interactive video tools on personal computers and
workstations, program delivery using cable and satellite,
and high-definition television (HDTV) are available for
visual communications. Many of these applications,
however, require the use of data compression because visual
signals require a large communication bandwidth for
transmission and a large amounts of computer memory for
storage [2][3]. In order to make the handling of visual
signals cost effective it is important that their bandwidth be
compressed as much as possible. Fortunately, visual signals
contain a large a mount of statistically and psychovisually
redundant information [4]. By removing this unnecessary
information, the amount of data necessary to adequately
represent an image can be reduced.
The removal of unnecessary information generally can
be achieved by using either statistical compression
techniques or psychovisual compression techniques. Both
techniques result in a loss information, but in the former the
loss may be recovered by signal processing such as filtering
and inter or intra-polation. In the later, information is in fact
discarded, but in way that is not perceptible to a human
observer. The later techniques offer much greater levels of
ISSN: 2085-6350
compression but it is no longer possible to perfectly
reconstructed the original image [4]. While the aim in
psychovisual coding is to keep these differences at an
imperceptible level, psychovisual compression inevitably
involves a tradeoff between the quality of the reconstructed
image and the compression rate achieved. This tradeoff can
often be assessed using mathematical criteria, although a
better assessment is in general provided by human observer.
The applications of image data compression, in general,
are primarily in the transmission and storage of information.
In transmission, applications such as broadcast television,
teleconferencing, videophone, computer-communication,
remote sensing via satellite or aircraft, etc., require the
compression techniques to be constrained by the need
For the real time compression and on-line consideration
which tends to severely limit the size and hardware
complexity. In storage applications such as medical images,
educational and business documents, etc., the requirements
are less stringent because much of the compression
processing can be done off-line. However, he
decompression or retrieval should still be quick and
efficient to minimize the response time [5].
All images of interest usually contain a considerable
amount of statistically and subjectively superfluous
information [6]. A statistical image compression technique
exploits statistical redundancies in the information in the
image. This technique reduces the amount of data to be
transmitted or to be stored in an image without any
information being lost. The alternative is to discard the
subjective redundancies in an image, which leads to
psychovisual image compression. These psychovisual
techniques rely on properties of the Human Visual
characteristic system (HVS) to be determined which
features will not be noticed by human observer.
There are numerous way to achieve compression in
statistical image compression techniques such as Pulse
Code Modulation (PCM), Differential PCM (DPCM),
Predictive Coding, Transform Coding, Pyramid Coding and
Subband Coding, as well as Psychovisual Coding
techniques. Statistical compression techniques all use a
form of amplitude quantization in their algorithms to
improve compression performance. Simple quantization
alone, however, is not the most efficient or flexible
techniques to combine with a statistical compression
algorithm [7]. The combination of quantization and
psychophysics, on the other hand has the potential to
remove most subjectively redundant information efficiently
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
Keynote
from an image, a process which is based on the actual
behavior of the HVS. Furthermore, subband coding and
pyramid coding schemes can be combined with visual
psychophysics-based compression techniques, since both of
these statistical schemes break the original image data down
into separate frequency bands. A process that similar to the
bandpass filter characteristics of the HVS and can be used
to quantized the information in each band depending on the
relative frequency band.
Transform coding is able to achieve optimum statistical
compression ratios, especially the Discrete Cosine
Transform (DCT). Much research has been performed in
combining the DCT transform coding and visual
psychophysics-based compression techniques [8][9][10]
resulting in a higher compression ratio and good
reconstruction of the original image.
Image compression techniques mentioned above,
involve spatial correlations in single frames where
redundancies are exploited either statistically or
subjectively, are known as intraframe coding techniques.
Interframe coding techniques known video compression, by
contrast, attempt to exploit the redundancies produced by
temporal correlation as well as spatial correlations in
successive video signals. These techniques hold the promise
of significantly greater reduction in the data required to
transmit the video signal as compared to interframe coding.
The simplest interframe coding technique is called
“conditional replenishment [11][12][13]. This technique
bases the coding scheme on the previous frame and is also
often called predictive coding. In the conditional
replenishment technique, only pixels the values of which
have changed significantly since the last frame, as
compared to a certain threshold, are transmitted. Another
technique, which still uses predictive coding from previous
frame, is adaptive intra-inter-frame prediction [14]. In this
technique, interframe prediction is used for scenes in
images where there is little motion, while intraframe
prediction is used for areas where this is much motion. The
switching between intra- and inter-frame prediction or a
combination of both, is usually controlled by the signal
changes of previously transmitted pixels so that no
overhead control information need to sent. The prediction
error can be quantized and transmitted for every pixel or can
be thresholded into predictable and non-predictable pixels
[15].
Adaptive prediction displacement of a moving object
which is based on information obtained from successive
frames is known as Motion Compensation. This scheme
was studied by [16] and [17] by measuring small
displacements based on very simple model of moving
objects in a stationary background scene for segmentation
purposes. A later refinement developed by [18][19][20] led
to one set of techniques known as Pel Recursive Motion
Compensation, which recursively adjusts the translational
estimates at every pixel or every small block of pixels. [21]
developed another technique known Block Matching
Motion Estimation. This technique estimates the location of
a block of pixels in the current frame by using a search in a
-5
confined window defined in the previous frame. Location of
the block results in the displacement vector for that block.
Different search methods have been proposed to avoid an
exhaustive search [22][23][24][25].
In order to produce a higher compression ratio,
transform coding has been applied to video coding, and can
be carried out as a three-dimensional transform [26] or in an
interframe coding scheme [27][28]. In the latter case motion
compensation can be performed in either the spatial domain
or the frequency domain. Transform coding can also be
combined with predictive coding so that the transform
coefficients from intraframe transformations of the previous
frame can be used to predict the transform coefficients of
the current frame [29]. CCITT H.261 Recommendations
[30], JPEG standard [31] and the MPEG draft [32], are also
DCT transform based and intra-inter-frame adaptive with
optional motion compensation. Their schemes result in a
blocking effect for low bit rates. Wavelet transform coding
can
effectively
eliminate
this
blocking
effect
[33][34][35][36][37] because the wavelet bases in adjacent
subbands overlap one another. Another advantage of
wavelet transform coding is that it is very similar to
subband coding. Wavelet transform combined by
psychovisual coding resulted a very good performance in
term of compression ratio and reconstructed images [4].
Since then, the DCT is replaced to the WT in order to
achieve high compression algorithms and good quality
reconstructed images, and has been adopted to be standard.
These standards are ITU standard for H-261, H-263, H-264;
ISO/IEC for JPEG, JPEG-2000, MPEG-2, MPEG-4, and
MPEG-7.
However, there is inevitably space for
improvements and extension within this area of research,
such as a hybrid system by using combining transform
method and Fuzzy, Neural Network, etc. For instance, the
TEMATICS Team has been developed some algorithm and
practical for analysis and modeling of video sequences;
sparse representations, compression and interaction with
indexing; Distributed source & Joint Source-Channel
Coding, etc [38].
References:
[1]
Li Deng, “Embracing A new Golden Age of Signal
Processing”, IEEE Signal Processing, Jan., 2009.
[2]
Dadang Gunawan, “Interframe Coding and Wavelet
Transform”, Journal IEICE, Vol. 1, No 1, pp. 22 – 37, Oct.,
1999.
[3]
Dadang Gunawan, “From Image to Video Compression”,
Jurnal Teknologi, Vol. IX, No. 2, Sep., 1995.
[4]
Dadang Gunawan & D.T. Nguyen, “Psychovisual Image
Coding using Wavelet Transform”, Australian Journal of
Intelligent Information Processing Systems, Autumn
Issues, Vol. 2, No. 1, Mar.,1995.
[5]
A. K. Jain, “Image Data Compression”, Proceeding of the
IEEE, Vol. 69., pp. 349 – 389, Mar., 1981.
[6]
Arun N Netravali & Barry G Haskell’ “Digital Pictures :
Representation and Compression”, Plenum Press, new
York, 1988.
[7]
David L McLarent, “Video and Image Coding for
broadband ISDN”, Ph.D. Thesis, University of Tasmania,
Australia, 1992.
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
6
[8]
[9]
[10]
[11]
[12]
[13]
[14]
[15]
[16]
[17]
[18]
[19]
[20]
[21]
[22]
Keynote
K. N. Ngan, K. S. Leong & H. Singh, “Adaptive Cosine
Transform Coding of Images in perceptual Domain”, IEEE
Transaction ASSP, Vo. 37, pp. 1743 – 1750, Nov., 1989.
B Chitprasert & K. R. Rao, “Human Visual Weighted
progressive Image Transmission”, IEEE Transaction on
Communication, Vol. 38, pp. 1040 – 1044, Jul. 1990.
D. L. McLaren & D. T. Nguyen, “The Removal Subjective
redundancy fro DCT Coded Images”, IEE Proceeding –
Part I, Vol. 138, pp. 345 – 350, Oct. 1991.
F. W. Mounts, “A Video Coding System with Conditional
Picture-Element Replenishment”, The Bell System
Technical Journal, Vol. 48, pp. 2545 – 2554, Sep. 1969.
J. C. Candy, “Transmitting television as Clusters of Frame
to frame Differences”, The Bell System Technical Journal,
Vol. 50, pp. 1889 – 1917, Aug. 1971.
[B. G. Haskel, F. W. Mount and C. Candy, “Interframe
Coding of Videotelephone Pictures”, Proceeding of IEEE,
Vol. 60, pp. 792 – 800, Jul. 1972.
D. Westerkamp, “Adaptive Intra-Inter frame DPCMCoding for Transmission TV-Signals with 34 Mbps”, IEEE
Zurich Seminar on digital Communication, pp. 39 – 45,
Mar. 1984.
M. H. Chan, “Image coding Algorithms for Videoconferencing Applications”, Ph.D. Thesis Imperial College
– University of London, 1989.
J. O. Limb & J. A. Murphy, “Measuring the Speed of
Moving Objects from television Signals”, IEEE
Transaction on Communication, Vol. 23, pp. 474 – 478,
Apr. 1975.
C. Cafforio & F Rocca, “Methods of Measuring Small
Displacements of Television Images”, IEEE Transaction
on Information Theory, Vol. 22, pp.573 – 579, Sep. 1976.
A. N. Netravali & J. D. Robbins, “Motion Compensated
television Coding ; Part 1”, The Bell-System Technical
Journal, Vol. 58, pp. 631 – 670, Mar. 1979.
C. Cafforio & F Rocca, “The Differential method for
Motion Estimation”, Image Science Processing & Dynamic
Scene Analysis, Springer Verlag, New York, pp. 104 –
124, 1983.
J. D. Robbins & A. N. Netravali, “Recursive motion
compensation : A Review”, mage Science Processing &
Dynamic Scene Analysis, Springer Verlag, New York, pp.
75, 1983.
J. R. Jain & A. K. Jain, “Displacement measurement & Its
Application in Interframe Image Coding”, IEEE
Transaction on Communication, Vol. 29, pp. 1799 – 1808,
Dec. 1981.
T. Koga, K.Iinuma, A. Hirano, Y. Iijima & T. Ishiguro,
“Motion-compensated Interframe Coding for Video
Conferencing”, Proceeding National Telecommunication
Conference, New Orleans, LA., pp. G5.3.1 – 5.3.5, Nov.
1981.
ISSN: 2085-6350
[23]
[24]
[25]
[26]
[27]
[28]
[29]
[30]
[31]
[32]
[33]
[34]
[35]
[36]
[37]
[38]
Proceedings of CITEE, August 4, 2009
R. Srinivasan & K. R. Rao, “Predictive Coding Based on
Efficient Motion Compensation”, IEEE International
Conference on Communication, Amsterdam, pp. 521 –
526, May 1984.
A. Puri, H. M. Hang & D. L. Schilling, “An Efficient
Block matching Algorithm for Motion Compensated
Coding”, Proceeding IEEE ICASSP, pp. 25.4.1 – 25.4.4,
1987.
M. Ghanbari, “The Cross Search Algorithm for Motion
Compensation”, IEEE Transaction on Communication,
Vol. 38, pp. 950 – 953, Jul. 1990.
M. Gotze & G Ocylock, “An Adaptive Interframe
Transform Coding System for Images”, proceeding IEEE
ICASSP 82, pp. 448 – 451, 1982.
J. R. Jain & A. K. Jain, “Displacement measurement & Its
Application in Interframe Image Coding”, IEEE
Transaction on Communication, Vol. 29, pp. 1799 – 1808,
Dec. 1981.
J. A. Roese, W. K. Pratt & G. S. Robinson, “Interframe
Cosine Transform Image Coding, “ IEEE Transaction on
Communication, Vol. 25, pp. 1329 – 1338, Nov. 1977.
J. A. Roese, “Hybrid Transform predictive Image Coding
in Image Transmission Techniques, Academic Press, new
York, 1979.
CCITT H.261 Recommendations, “Video Codec for
Audiovisual Services at p x 64 kbps”, 1989.
G. Wallace, “The JPEG Stil Picture Compression
Standard”, Communication ACM, Vol. 34, pp. 30 – 44,
Apr. 1991.
D. LeGall, “MPEG : A Video Compression Standard for
Multimedia Applications”, Communication ACM, Vol. 34,
pp. 46 – 58, Apr. 1991.
S. G. Mallat, “A Theory for Multiresolution Signal
Decomposition : the Wavelet Representation” IEEE
Transaction on Pattern Analysis & Machine Intelligent,
Vol. 11, pp. 674 – 693, Jul. 1989.
S. G. Mallat, “Multifrequency Channel Decomposition of
Image and Wavelet Models”, IEEE Transaction on ASSP,
Vol. 37, pp. 2091 – 2110, Dec. 1989.
O Riol & M. Vetterli, “Wavelet & Signal Processing”,
IEEE Signal Processing Magazine, Vol. 8, pp. 14 – 38,
Oct. 1991.
Y. Q. Zhang & S. Zafar, “Motion-Compensated Wavelet
Transform Coding for Color Video Compression”, IEEE
Transaction on Circuit & Systems for Video Technology,
Vol. 2, pp. 285 – 296, Sep. 1992.
S. Zafar, Y. Q. Zhang & B. Jabbari, “Multiscale Video
Representation Using Multiresolution Compensation &
Wavelet Decompostion“, IEEE Journal Selected Area in
Communications, Vol. 11, pp. 24 – 34, Jan. 1993.
Project Team Tematics, “Activity Report”, INRIA, 2008.
Conference on Information Technology and Electrical Engineering (CITEE)
7 Proceedings of CITEE, August 4, 2009
Analysis and Simulation Of Bus-Clamping
PWM Techniques Based On Space Vector
Approach
1
2
Ms.M.Thanuja1
Mrs. K. SreeGowri2
PG-Student, Dept. of Electrical and Electronics, RGMCET, Nandyal, India.
Assoc. Professor, Dept of Electrical and Electronics, RGMCET, Nandyal, India.
Abstract:
Conventional space vector pulse width modulation
employs conventional switching sequence, which divides
the zero vector time equally between the two zero states
in every subcycle.Existing bus-clamping PWM
techniques employ clamping sequences, which use only
one zero state in a subcycle.In the present work a new
set of BCPWM dealing with a special type of switching
sequences, termed as “double-switching
clamping
sequences”, which use only one zero state and an active
vector repeats twice in a sub cycle, will be proposed.
It is shown analytically that the proposed
BCPWM techniques result in reduced harmonic
distortion in the line currents over CSVPWM as well as
existing BCPWM techniques at high modulation indices
for given a average switching frequency. This work
deals with Analysis and Simulation of “double-switching
clamping sequences” in terms of stator flux ripple and
line current harmonic distortion. Simulation is done on
v/f
controlled
Induction
Motor
drive
in
MATLAB/SIMULINK environment.
Index Terms—Bus clamping pulse width modulation
(BCPWM),discontinuous PWM, harmonic distortion,
induction motor drives,PWM inverters, space vector
PWM, stator flux ripple, switching sequences.
v
I. INTRODUCTION:
oltage source inverter fed induction motors are widely
used in variable speed applications. The harmonic
distortion in the motor phase currents must be low for
satisfactory operation of
the
motor drive.
The
harmonic
distortion in the current is determined by the switching
frequency and PWM Technique is employed. The switching
frequency cannot be increased beyond a certain range due to
practical limitations. The distortion is reduced at a given
switching frequency by a good design of PWM Technique.
Specially designed PWM Techniques have been
reported for high power applications where the inerter
switching frequency is low. For switching frequency much
higher than the maximum fundamental frequency, several
modulation functions and frequency modulation of carrier
have been investigated. This project focuses on developing
and evaluating new real time PWM techniques for voltage
source inverters.
SPWM and CSVPWM are very popular real time
techniques. CSVPWM and THIPWM lead to higher line
side voltages for given dc bus voltage compare to SPWM.
These technique results in less harmonic distortion in motor
currents than SPWM at a given line voltage. Discontinuous
modulation methods lead to reduction in distortion at higher
line voltages over a CSVPWM for a given average
switching frequency. This paper proposes high performance
HSVPWM, which further reduce the distortion in the line
currents over comparable real-time technique at a given
average switching frequency. The superiority in
performance of proposed techniques is established
theoretically as well as experimentally.
With SPWM, CSVPWM and THIPWM, every phase
switches once in a sub-cycle or half carrier signal. This
paper explores novel switching sequence that switch ‘a’
phase twice in a sub-cycle, while switching second phase
once and clamping the third phase. This paper brings out all
such possible sequences (including two new sequences),
which results same average switching frequency as
CSVPWM for a given sampling frequency.
Real-time PWM techniques balance the reference
volt-seconds and applied volt-second over every sub cycle
or half carrier cycle. The multiplicity of possible switching
sequences provides a choice in the selection of switching
sequences in every sub cycle. The proposed hybrid PWM
techniques employ the sequence which results in the lowest
rms current ripples over given subcycle, out of given set of
sequences. Consequently the total rms current ripple over
fundamental cycle is reduced
II.SWITCHING SEQUENCES OF INVERTER Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350 Proceedings of CITEE, August 4, 2009 8
duration for which the active state is applied in a sub cycle
must satisfy (1).
4) The total duration for which the zero vector (either using
the zero state 0 or the zero state 7) is applied in a sub cycle
must satisfy (1).
5) Only one phase must switch for a state transition.
6) The total number of switching’s in a subcycle must be
less
than or equal to three. This ensures that the average
switching frequency is less than or equal to that of
CSVPWM for a given sampling Frequency.
Fig.1. Two level inverter circuit diagram
A three-phase voltage source inverter has eight switching
states as shown in Fig. 1. The two zero states (−−− and
+++), which short the motor terminals, produce a voltage
vector of zero magnitude as shown in the figure. The other
six states, or the active states, produce an active voltage
vector
each. These active vectors divide the space vector plane into
six sectors and are of equal magnitude as shown. The
magnitudes are normalized with respect to the dc bus
voltage.
From the Volt-time balance principle T1, T2 and Tz can be
given as
T1 = Vref * Ts *
sin (60° − α )
sin (60°)
T2 = Vref * Ts *
sin (α )
sin (60°)
Tz = Ts − T1 − T2
(2)
Fig.2. Inverter states and voltage vectors of three-phase
Inverter
The six active space vectors are represented by the
following expression:
(1)
In space vector-based PWM, the voltage reference, which is
sampled once in every subcycle, TS. Given a sampled
reference vector of magnitude VREF and angle α in sector I
as shown in Fig. 2, the dwell times of active vector 1, active
vector 2 and zero vector in the subcycle are given by T1, T2,
and TZ, respectively, in CSVPWM divides TZ equally
between 0 and 7, and employs the switching sequence 0-12-7 or 7-2-1-0 in a subcycle in sector I. The conditions to be
satisfied by a valid sequence in sector I are as follows.
1) The active state 1 and the active state 2 must be applied
at least once in a subcycle.
2) Either the zero state 0 or the zero state 7 must be applied
at least once in a subcycle.
3) In case of multiple application of an active state, the total
ISSN: 2085-6350
Fig.3. Different possible switching sequences in sector I
Conference on Information Technology and Electrical Engineering (CITEE) 9 Proceedings of CITEE, August 4, 2009
Fig.4. PWM Gate signals when the reference vector
sitting in sector-I (0127)
Sequence 7212 leads to clamping of R -phase to the
positive dc bus, while sequence 0121 results in clamping of
B-phase to the negative dc bus. Both sequences result in Y phase switching twice in a sub cycle. Hence, sequences
0121 and 7212 are termed “double-switching clamping
sequences” here. The sequences illustrated in Fig. 3 are
employed in sector I. The equivalent sequences in the other
sectors are as listed in Table I. The PWM gating signals for
the CSVPWM is shown in Fig 4.
number of DPWM methods. If
δ = 0, -pi/6, -pi/3, then
DPWM1, DPWM2 and DPWM3 can be obtained
respectively. The modulation waveforms of the different
PWM methods are as shown in Fig.5. 1
TABLE I. SWITCHING SEQUENCES IN SIX SECTORS
III. MODERN PWM TECHNIQUES
The modern PWM methods can be
separated into two groups and those are:
•
•
Continuous PWM(CPWM) methods
Discontinuous PWM(DPWM) methods
In discontinuous one the modulation wave of a phase has
at least one segment which is clamped to the positive or
negative dc bus for at most a total of 1200 (over a
fundamental cycle).Where as in continuous PWM there is
no clamping in the modulation wave.
The expressions for the modulation signals are
given as
M in =
Fig.5.Modulation waveforms of the various PWM
methods
The conventional SVPWM algorithm employs equal
division of zero voltage vector times within a sampling
period or sub cycle. However, by utilizing the freedom of
zero state division, various DPWM methods can be
generated. GDPWM algorithm, which
uses the utilization of the freedom of zero state time
division. In this proposed method the zero state time will be
shared between two zero states as T0 for V0 and T7 for V7
respectively, and T0 , T7 can be expressed as:
2V in
2 µ V min 2 (µ − 1)V max
+ (1 − 2µ ) −
+
Vdc
Vdc
Vdc
i=a, b, c
(5)
(4)
The selection of µ gives rise to an infinite number of PWM
modulations. To obtain the generalized discontinuous
modulation signal, µ is given as :
μ =1-0.5[1+sgn (cos3 (ωt+δ))]
(3)
When µ = 0, any one of the phases is clamped to
positive dc bus for 1200 and then DPWMMAX is obtained.
When µ = 1, any one of the phases is clamped to negative
dc bus for 1200 and then DPWMMIN is obtained. If µ = 0.5,
then the SVPWM algorithm is obtained. Similarly, the
variation of modulation phase angle δ yields to infinite
Fig.6.Existing
clamp).
Conference on Information Technology and Electrical Engineering (CITEE)
bus-clamping
PWM
technique
(300
ISSN: 2085-6350 Proceedings of CITEE, August 4, 2009 10
IV. BUS-CLAMPING PWM TECHNIQUES
A popular existing bus-clamping method clamps
every phase during the middle 300 duration in every quarter
cycle of its fundamental voltage. This technique, termed as
and “300 clamp.” This employs sequences 721, 127, in the
first half, and 012, 210, in the second half of sector I as
shown in Fig. 6.
(6)
In Fig. 7(a), every phase is clamped
continually for 600 duration in every half cycle of the
fundamental voltage waveform. These techniques can be
termed “continual clamping” techniques. In Fig. 7(b), the
600 clamping duration is split into one interval of width in
the first quarter cycle and another interval of (600-γ) in the
next quarter in every half cycle. Since the clamping duration
is split into two intervals, these techniques are termed “split
clamping” PWM techniques. Fig. 8(a) and (b) present
average pole voltage waveforms that illustrate the two types
of clamping for γ= 450 .
Fig.9.Simulated Phase and line voltages waveforms of
the two level inverter
VOLTS /HZ CONTROL TECHNIQUE: Fig.7. Existing bus-clamping PWM techniques
(a) continual clamping type (b) split clamping type.
This is the most popular method of Speed control
because of simplicity. The Flux and Torque are also
function of frequency and voltage respectively the
magnitude variation of control variables only. The air gap
voltage of induction motor is
φag
=
Eag
f
≈
V
f
Speed is varied by varying the frequency; maintain v/f
constant to avoid saturation of flux. With constant v/f ratio,
motor develops a constant maximum torque.
Fig.8. Average pole voltage over a fundamental cycle for
VREF = 0.75 corresponding to (a) continual clamping
and (b) split clamping both with γ=450.
The design of the inverter phase voltage and common mode
voltages for different pulse sequences are:
ISSN: 2085-6350
INDUCTION MOTOR MODELLING:
Among the various reference frames, V/F uses the
stationary reference frame. Hence, in this work, the
induction motor model is developed in the stationary
reference frame, which is also known as Stanley reference
frame. Rotor and stator voltages and their flux linkages are
Conference on Information Technology and Electrical Engineering (CITEE) 11 Proceedings of CITEE, August 4, 2009
vds = Rs ids +
vqs = Rs iqs +
dψ ds
dt
dψ qs
ψ qs = Ls iqs + Lm iqr
dt
vdr = Rr idr + ω rψ qr +
vqr = Rr iqr − ω rψ dr +
dψ dr
dt
dψ qr
ψ ds = Ls ids + Lm idr
ψ qr = Lr iqr + Lm iqs
ψ dr = Lr idr + Lm ids
dt
The electromagnetic torque of the induction motor is given
by
dω m
2 dω r
= TL + J
Te = TL + J
dt
P dt
Q1 = [cos (α) −Vref] *T1 Q2 = [cos (60° − α) − Vref] *T2
QZ = −Vref*TZ
D = sin (α)*T1.
Expressions for RMS Stator Flux Ripple:
The rms Stator flux ripples different sequences employed
and their respective vector diagram of d-axis and q-axis
ripples shown in Fig 12.
[
]
1
1
2 T
2
2 T
2
F0127
= (0.5Qz ) z + (0.5Qz ) + 0.5Qz (0.5Qz +Q1) + (0.5Qz + Q1) 1 +
3
2Ts 3
Ts
[
]
1
(0.5Qz +Q1)2 −(0.5Qz +Q1)(0.5Qz ) +(−0.5Qz )2 T2 +
3
Ts
The Electromechanical equation of induction drive is given
by
1
(−0.5Qz )2 Tz + 1 D2 (T1 +T2 )
3
2Ts 3
Ts
3⎛P⎞
T e= ⎜ ⎟(ψ ds iqs −ψ qs ids )
2⎝ 2 ⎠
(7a)
[
]
1 T 1
2 T
2
F012
= Qz2 z + Qz2 + Qz (Qz + Q1 ) + (Qz + Q1 ) 1
3 Ts 3
Ts
1
1 (T + T )
2T
+ (Qz + Q1 ) 2 + D2 1 2
3
Ts 3
Ts
(7b)
[
]
1 T 1
2 T
2
= Qz2 z + Qz2 + Qz (Qz + Q2 ) + (Qz + Q2 ) 2
F721
3 Ts 3
Ts
+
Fig.10. Speed and Torque characteristics with respect to
time
1
(Qz + Q2 )2 T1 + 1 D2 (T1 + T2 )
3
Ts 3
Ts
(7c)
[
]
1 T 1
2 T
2
F7212
= Qz2 z + Qz2 + Qz (Qz + 0.5Q2 ) + (Qz + 0.5Q2 ) 2 +
3 Ts 3
2Ts
[
]
1
(Qz + 0.5Q2 )2 − (Qz + 0.5Q2 )0.5Q2 + (− 0.5Q2 )2 T1 +
Ts
3
1
(− 0.5Q2 )2 T2 + 1 (0.5D)2 (T1 + T2 )
Ts
3
2Ts 3
.
.
Fig.11. Block diagram of V/F controlled
based IM drive
2
=
F0121
BCPWM
+
V. ANALYSIS OF HARMONIC DISTORTION
The generalized stator q-axis and d-axis flux ripples are
as shown below.
(7d)
+
[
]
1 2 Tz 1 2
2 T1
+ Q z + Q z (Q z + 0.5Q1 ) + (Q z + 0.5Q1 )
Qz
3
2Ts
Ts 3
[
]
1
(Q z + 0.5Q1 )2 − (Q z + 0.5Q1 )0.5Q1 + (− 0.5Q1 )2 T2
3
Ts
1
(− 0.5Q1 )2 T1 + 1 (0.5 D )2 (T1 + T2 )
3
2Ts 3
Ts
.
Conference on Information Technology and Electrical Engineering (CITEE)
(7e)
ISSN: 2085-6350 Proceedings of CITEE, August 4, 2009 12
Fig. 13. Comparison of rms stator flux ripples due to
CS-0127, S1-012, S2-721, S3-0121, S4-7212, S5-1012 and
S6-2721 at different modulation indices.
A. Analysis of Existing BCPWM Techniques:
Sequence 012 leads to less RMS current ripple over
a subcycle than 721 in the first half of the sector, and vice
versa in the second half of the sector.
F012(α) < F721(α) , 00 < α < 300
(8a)
F012(α) > F721(α) , 300 < α < 600
(8b)
F012(α) = F721(600- α)
(9)
Fig.12.Stator flux ripple vector over a subcycle for
sequences (a) 0127, (b) 012, (c) 721, (d) 0121 and (e)
7212.
Fig.14.Measured no-load current waveforms at
VREF=0.85 for Existing BCPWM techniques.
ISSN: 2085-6350
Conference on Information Technology and Electrical Engineering (CITEE) 13 Proceedings of CITEE, August 4, 2009
techniques. The switching energy loss in a subcycle in an
inverter leg is proportional to the phase current and the
number of switchings of the phase (n) in the given subcycle.
The normalized switching energy loss per subcycle (ESUB)
in an inverter leg is defined in (1a), where i1 is the
fundamental phase current, Im is the peak phase
fundamental current and Φ is the line-side power factor
angle. Table II Measured Values of ITHD for Existing
BCPWM
No-Load Current THD
γ=300
γ =450
continual clamping
5.32%
4.72%
Split Clamping
4.61%
5.01%
E SUB =
n i1
Im
= n sin (ωt − Φ )
(12a) Π
E SUB (AV ) =
1
E SUB dωt
Π ∫0
(12b)
Fig.15.Measured no-load current waveforms
VREF=0.85 for Proposed BCPWM techniques.
at
Fig.16. Variation of normalized switching loss ESUB over
a fundamental cycle for CSVPWM.
B. Analysis of Proposed BCPWM Techniques:
F0121(α) < F7212(α) , 00 < α < 300
(10a)
F0121(α) > F7212(α) , 300 < α < 600
(10b)
F0121(α) = F7212(600- α)
(11)
Table III Measured Values of ITHD for Proposed
BCPWM
No-Load Current THD
γ=300
γ=450
continual clamping
3.91%
3.46%
Split Clamping
2.66%
3.04%
VI INVERTER SWITCHING LOSSES
This section presents a comparison of inverter
switching losses due to CSVPWM, existing BCPWM
(a)
(b)
(c)
(d)
Fig. 17. Variation of normalized switching loss ESUB over
a fundamental cycle for Existing BCPWM techniques.
Average Switching Loss for CSVPWM = 0.6366
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350 Proceedings of CITEE, August 4, 2009 14
Table IV Measured Values of Average Switching Loss
Average Switching Loss
Φ=00
Φ =900
(a) 0.318
----
-----
(b) 0.4776
(c) 0.4041
(d) 0.4041
continual clamping
(i) γ=300
(ii)
γ=600
Split Clamping
γ=300
VI. CONCLUSIONS
A class of bus-clamping PWM (BCPWM)
techniques, which employ only the double-switching
clamping sequences, is proposed. The proposed BCPWM
techniques are studied, and are compared against
conventional space vector PWM (CSVPWM) and existing
BCPWM techniques at a given average switching
frequency. The proposed families of BCPWM techniques
result in less line current distortion than CSVPWM and the
existing BCPWM techniques at high line voltages close to
the highest line side voltage during linear modulation. The
analysis presented explains the difference in distortion due
to the different techniques. The study classifies both the
existing BCPWM and the proposed BCPWM techniques
into two categories, namely continual clamping methods
and split clamping methods, depending on the type of
clamping adopted. It is shown that split clamping methods
are better than continual clamping ones in terms of line
current distortion. In terms of switching losses, continual
clamping is better at high power factors, while split
clamping is superior at low power factors. REFERENCES
[1]“Advanced Bus-Clamping PWM Techniques Based on
Space Vector Approach” G. Narayanan, Member, IEEE,
ISSN: 2085-6350
Harish K. Krishnamurthy, Di Zhao, and Rajapandian
Ayyanar, Member, IEEE,2006.
[2] J. Holtz, “Pulse width modulation—A survey,” IEEE
Trans Ind. Electron., vol. 39, no. 5, pp. 410–420, Dec.
1992.
[3] J. Holtz, “Pulse width modulation for electronic power
conversion,” Proc. IEEE, vol. 82, no. 8, pp. 1194–1214,
Aug. 1994.
[4] D. G. Holmes and T. A. Lipo, Pulse Width Modulation
for Power Converters: Principle and Practice. New York:
Wiley, 2003.
[5] V. Blasko, “Analysis of a hybrid PWM based on
modified space-vector and triangle-comparison methods,”
IEEE Trans.
Ind. Appl., vol. 33, no. 3, pp. 756–764,
May/Jun. 1997.
[6] D. Zhao, G. Narayanan, and R. Ayyanar, “Switching
loss characteristics of sequences involving active state
division space vector based PWM,” in Proc. IEEE
APEC’04, 2004, pp. 479–485.
ACKNOWLEDGEMENT
We express our sincere thanks to TEQIP for providing us good
lab facilities. A heart full and sincere gratitude to our beloved
parents for their tremendous motivation and moral support.
M.Thanuja was born in Nandyal. She has completed her
B.Tech at R.G.M. College of Engg and Techng. affiliated to
J.N.T.U. Hyderabad in 2002. She is currently doing M.Tech
at RGMCET,Nandyal, affiliated to J.N.T.U. Anantapur. She
is working in areas power electronics and Power electronics
control of ac drives.
E-mail: [email protected]
K.SreeGowri received the B.Tech degree from SVU
College of Engineering, Tirupati in 1997, the M.Tech
degree from RGM College of Engineering and Technology,
Nandyal and is currently pursuing the Ph.D. in Electrical
Engineering Department, JNTU, and Hyderabad. She is
currently an Associate Professor in the Department of EEE
in RGMCET, Nandyal, A.P. Her areas of interest include
Power Electronics, pulse width modulation techniques, AC
Drives and Control.
Conference on Information Technology and Electrical Engineering (CITEE) Proceedings of CITEE, August 4, 2009
15
Medical Image Processing of Proximal Femur
X-Ray for Osteoporosis Detection
Riandini, Mera Kartika Delimayanti
Politeknik Negeri Jakarta, Electrical Engineering Department
Kampus UI Depok, Indonesia 16424
[email protected]
Donny Danudirdjo
Institut Teknologi Bandung
Jl. Ganesha 10 Bandung, Indonesia 40132
Abstract— Osteoporosis is one of the generative (aging
process) diseases which occurs because of the rise of one’s life
expectancy from the age of 60s to 74s or more. This disease
attacks not only women but also men. The worst impact of it is
death. Although it is very important to detect it very early to
avoid the worst impact of it, an early effective process of
osteoporosis detection is clinically difficult and costly to do.
The only method of doing so which is relatively cheap is the
analysis of x-ray image (radiology) instead of DEXA (Dual
Energi X-Ray Absorptiometri) which long has been used as a
golden standard. But, this process may lead to subjectivity
since doctors (medical staffs) do so with naked eyes. This paper
describes a series of algorithm plans for medical image
analysis with the support of Femur Proximal x-ray image as
data input. Algorithm plans which are going to be developed
are a combination of Gabor algorithm, Wavelet algorithm,
Fractal algorithm and algorithm (analysis) mainly used to
enhance the quality of an image through the omission of
vertical lines artefact on the image, edge detection on the
image and image rotation. The outcome of the plan is going to
be implemented within GUI-based (Graphical User Interface)
computer application which is user-friendly and usable for
helping doctors diagnose osteoporosis.
Keywords—medical image analysis, x-ray image, proximal
femur, osteoporosis, algoritma.
I. INTRODUCTION
Osteoporosis is a degenerative disease that takes place in
many places around the world. The high prevalence of
osteoporosis in Indonesia is inseparable from the rise of life
expectancy, from 59.8 in 1990 up to 67 at present. An
effective standard instrument largely used nowadays to
detect osteoporosis is Dual Energy X-Ray Absorptiometri
(DEXA). A significant shortcoming of the instrument is that
it is very expensive and the number of it is limited not only
at rural areas but also in big cities over Indonesia. Besides,
the operational expenditure of the tool is totally high. One
alternate solution to the problem is the implementation of xray radiology image detecting instrument. However, it still
brings two primary weaknesses: (1). Data processing
subjectivity – a physician analyzes and interprets the same
data several times with different results; (2) Data processing
with the naked eye may lead to ignorance of detailed
information on a radiology image. In order to deal with
those weaknesses, this paper is going to describe a system
of medical image processing algorithm with a purpose of
assisting doctors for early detection of osteoporosis. The
system applies femur proximal x-ray image in its operation.
The algorithm operation of image texture analysis involves
determination of optimum textural features of the image and
the supporting classification methods of the image texture.
The system used in the operation is then used by physicians
to assist them diagnose osteoporosis.
II. OSTEOPOROSIS
Osteoporosis is a situation when there is an over-limit
decrease of bone density and structural malfunction of bone
micro-structure system that makes the bone system unable
to prevent bone fracture from minimal trauma. The decrease
of bone mass occurs because the velocity of bone formation
process by osteoblas cell is not capable of equalizing the
velocity of bone surface erodibility by osteoklas cell.
Histopalogically, osteoporosis indicates the decrease of
cortex density and the declining amount as well as size of
bone trabucela [3].
Osteoporosis is suffered much more by women than men
in that about 50 percent of menopause women are suffering
from osteoporosis. However, recent research has shown that
it is no longer dominated by old-aged people. Youngsters
are likely to get suffered from it if they don’t have
appropriate eating habit [6].
The number of osteoporosis cases in Indonesia is quite
high. Prevalence of it has reached 19.7 percent. This
percentage is believed to keep increasing due to the rising
life expectancy and the improvement of the quality of
people’s health. Osteoporosis is categorized as a
degenerative disease which is commonly suffered by old
people and may lead to physical disability and death. Thus,
the early prevention of it is very crucial.
Efforts to prevent osteoporosis are closely related to the
early bone strength diagnosis and treatment. The standard
method of diagnosing the bone density is well known as
DEXA (Dual Energy X-Ray Absorptiometry). In Indonesia,
the application of the method faces two problems: (1) fairly
expensive diagnosis and treatment, and (2) lack of medical
instruments supporting the diagnosis process within some
hospitals in big cities of the country. An alternate method to
cope with the problem, which is aimed at doing trabekula
style treatment, is Singh Index.
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
16
Proceedings of CITEE, August 4, 2009
Fig 2.1. illustrates the way of determining Singh Index on
femur x-ray image
Index value determination is commonly done by an
orthopedic specialist. The outcome of the index value is
strongly influenced by a specialist expertise and inseparable
from his/her subjectivity. If he/she is doing the evaluation at
two distinct times, he/she may come up with different
outcomes. Therefore, the presence of the diagnosing system
of trabekula texture of the femur x-ray image is in urgent
need to help doctors do early osteoporosis diagnosis and
detection.
III. TEXTURE ANALYSIS Texture is a two dimension intrinsic character of an image
correlated with the level of roughness, granularity, and the
regularity of pixel arrangement. Texture analysis runs
through observing neighboring model within pixels in a
spatial domain and is related to the need of the image
characters extraction. In this research, an observation of
textural characteristics of x-ray radiology image (connected
to level of osteoporosis) is carried out with the support of
four methods: co-occurrence analysis, Gabor filter, wavelet
transform, and fractal analysis. These four methods are used
in an attempt to obtain the most optimum method and result
in an accurate diagnosis outcome. There are eight image
characters used in the research, and in each method involves
at least one character. Character classification process
works through the support of artificial neuron system with
an outcome of Singh index value of the radiology image
that is observed.
A.
Co­occurrence Analysis Co-occurrence analysis is the simplest analyzing
method of the image texture. This analysis is completed
ISSN: 2085-6350
through two stages: (1) establishing co-occurrence matrix as
a representation of the probability of neighboring
interaction between two pixels in certain distance (d) and
angle orientation (θ), (2) determining character as a function
of the co-occurrence matrix.
Co-occurrence is two concurrent events, i.e., the
number of the occurrence of one level pixel value as a
neighbor of another one level pixel value in certain distance
(d) and angle orientation (θ). Distance is valued in Pixel and
orientation in Degree. There are four angle orientations with
45° angle interval, which are 0°, 45°, 90°, and 135°. On the
other hand, distance within pixels is valued 1 pixel.
Co-occurrence matrix is a square matrix with the
number of elements equal to a square of the number of pixel
intensity on an image. Each spot (p, q) within a cooccurrence matrix oriented to θ has a possibility of pixel
occurrence with value amounting p that is neighboring with
another pixel with the value amounting q in distance d and
angle orientation θ and (180−θ). Figure 3.1 illustrates the
formation of a co-occurrence matrix.
With the outcomes of matrix co-occurrence above, we
can determine statistical character of orde 2 that represents
an observed image. Haralick proposed a number of textural
characters which are likely to be extracted from a matrix cooccurrence [1]. In this research, six textural characters of an
image are used: Angular Second Moment, Contrast,
Correlation, Variance, Inverse Difference Moment, and
Entropy.
1.) Angular Second Moment (ASM) This character is used to show the level of homogeneity of
an image
(1)
{p(i, j )}2
ASM =
∑∑
i
j
Where p(i,j) represents values of row i and column j on a
co-occurrence matrix.
2.) Contrast (CON) This character is used to show the spread (moment of
inertia) of all elements on an image matrix. If the spread
(location) is located far away from the main diagonal line
on the matrix, then it has a big contrast value. Visually,
contrast value is the level of greyness variedness on an
image.
⎡
⎤
CON = ∑ k 2 ⎢∑∑ p (i, j )⎥
k
⎣ i j
⎦
(2)
i− j = k
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
17
ENT2 = −∑∑ p (i, j ) ⋅ 2 log p (i, j )
i
(6)
j
B. Gabor Filter Human’s visualizing ability system to distinguish
different image textures is based on the ability to identify a
variety of spatial frequencies and orientations of an image’s
textures observed. Gabor Filter is one of the filters that can
represent human’s visualizing ability system in isolating
spatial frequencies and orientations of an image’s textures.
This character enables Gabor Filter to be used as an
application of an image texture observation [4].
Spatially, Gabor function is sinusoida function which is
modulated by Gaussian function. An impulse response of a
two-dimension complex Gabor filter is shown below:
⎧⎪ 1 ⎡ x 2 y 2 ⎤ ⎫⎪
(7)
1
h( x, y) =
Fig 3.1.Illustration of the matrix co-occurrence formation on an
image: (a) image input (b) intensity value of an image input (c) outcome of
0° co-occurrence matrix (d) outcome of 45° co-occurrence matrix (e)
outcome of 90° co-occurrence matrix (f) outcome of 135° co-occurrence
matrix
2πσ xσ y
exp⎨− ⎢ 2 + 2 ⎥ ⎬ exp( j 2πFx)
⎪⎩ 2 ⎢⎣σ x σ y ⎥⎦ ⎪⎭
Illustration of a Gabor filter within a spatial domain is
presented in Figure 3.2.
3.) Correlation (COR)
Correlation is used to show the level of linear interdependence of greyness on an image and thus indicates the
presence of linear structure on the image.
COR =
∑∑ (ij ). p(i, j ) − µ
i
x
µy
j
σ xσ y
Fig 3.2. Impulse Response of a two-dimension Gabor Filter
(3)
4.) Variance (VAR)
Variance is used to present variedness of elements of a cooccurrence matrix. An image with low degree of greyness
has a small value of variedness.
(4)
VAR =
(i − µ x )( j − µ y ) p(i, j )
∑∑
i
j
Within a spatial frequency domain, a Gabor filter is
represented
as
shown
below:
H (u, v) = exp{− 2π 2 (u − F ) 2 σ x2 + v 2σ y2 }
(8)
[
]
Illustration of a Gabor filter within a spatial frequency
domain is presented in Figure 3.3
5.) Inverse Different Moment (IDM)
Inverse Different Moment (IDM) is mainly used to show
homogeneity of an image with certain similar degree of
greyness. A homogeneous image will have a bigger IDM
value than a heterogonous image.
IDM = ∑∑
i
j
1
p(i, j )
1 + (i − j ) 2
(5)
6.) Entropy (ENT)
Entropy is used to show the level of shape irregularity. An
image with a big ENT value is the one that has even degree
of greyness; meanwhile, the image with a small ENT value
has uneven degree of greyness (varied degree of greyness).
Fig 3.3. Parameter of a Gabor Filter within a Spatial Frequency Domain
There are six parameters that must be present in the
implementation of a Gabor filter. They are F, θ, σx, σy, BF,
and Bθ .
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
18
Proceedings of CITEE, August 4, 2009
•
Frequency (F) and orientation (θ) define the
location of a filter center
•
BF and Bθ represent constants of the width of a
frequency tape and the angular range of a filter
•
Variabel σx is closely related to a response of -6 dB
for a spatial frequency component
•
Variabel σy is closely connected with a response of
-6dB for an angular component
σy =
•
ln 2
2πF tan( Bθ / 2)
Value used
Symbol
research
in
the
2 2 2 2 2
,
,
,
,
21 2 2 2 3 2 4 2 5
1 octave
Angular tape width
Bθ
45°
Frequency Spacing
SF
1 oktaf
Angular Spacing
Sθ
45°
Orientation
Θ
0°, 45°, 90°, 135°
N
∑∑ x(m, n)
2
k
⎣
J +1,k
J +1,k
∑
j =0
j +1,k
j +1,k
⎥
⎦
In which coefficient c0,k has been defined already and
coefficients cj+1,n and dj+1,n at scale j+1 are related to
coefficient cj,k at scale j through a formula:
c j +1,n = ∑ c j ,k h(k − 2n )
(14a)
d j +1,n = ∑ d j ,k g (k − 2n)
(14b)
k
In which equation 0 ≤ j ≤ J applies.
The main wavelet is established by scale function φ(t)
and wavelet ψ(t) which comply with two-scale relation as
shown in a below formula:
(15a)
φ ( x ) = 2 ∑ h(k ) φ (2 x − k )
ψ ( x ) = 2 ∑ g (k ) φ (2 x − k )
(15b)
k
With requirement:
k
g (k ) = (− 1) h(1 − k )
(10)
i =1 j =1
This research uses frequency tape width (BF) and middle
frequency distance (SF) of one octave, and angular tape
width (Bθ) and angular distance (Sθ) of 45°. These values are
highly likely to reflect characteristics of human’s
visualizing system.
C. Wavelet Transform Textural analysis with wavelet transform works through
decomposition of image input, used to observe the content
of image information within several sub-bands.
Decomposition process applying discrete wavelet transform
(DWT) is conducted in stages, each of which will result in 4
sub-band matrix – one matrix of approximation coefficient
A and three detailed matrix (horizontal H, vertical V, and
diagonal D). If the process is conducted in a narrower range
of frequencies, then a matrix of approximation coefficient A
is likely to be re-decomposed into 4 smaller sub-band
matrix [5].
ISSN: 2085-6350
∑⎢
0,k 0,k
k
After Gabor character is obtained, then extraction of energy
character is likely to carry out. This is defined as follow:
M
(12)
k
BF
1
MN
∑
k
Frequency tape width
E ( x) =
1 ⎛ x−b⎞
ψ⎜
⎟
a ⎝ a ⎠
ψ a ,b ( x ) =
0
Table 3.1. Six Parameters of a Gabor Filter
Middle
frequency
F
(normalized)
Where wavelet ψa,b is counted from the main wavelet
through translation and dilation
Wavelet decomposition at level J is written below
J
⎡
⎤ (13)
f (x) = c φ (x) = c φ (x) + d ψ (x)
(9)
Position (F, θ) and tape width (σx, σy) of a Gabor
filter within a frequency domain have to be
determined accurately to obtain correct textural
information. A middle frequency of a canal filter
must be close to a frequency of an image texture
characteristic.
Parameter
Continual wavelet transform for signal 1-D f(x) is
obviously defines as follow:
(11)
Wa f (b ) = ∫ f ( x )ψ a*,b ( x )dx
(15c)
Coefficient h(k) is one of the main wavelet functions
like Haar, Daubechies, Coiflets, Symlets, or Biorthogonal.
Practically, discrete wavelet transform is calculated by
applying a bank filter inseparable from function f(x) or
image function I(x).
Ln (bi , b j ) = [ H x * [ H y * Ln−1 ]↓2,1 ]↓1, 2 (bi , b j ) (16a)
Dn1 (bi , b j ) = [ H x * [G y * Ln −1 ]↓ 2,1 ]↓1, 2 (bi , b j ) (16b)
Dn 2 (bi , b j ) = [G x * [ H y * Ln−1 ]↓2,1 ]↓1, 2 (bi , b j ) (16c)
Dn3 (bi , b j ) = [G x * [G y * Ln −1 ]↓2,1 ]↓1, 2 (bi , b j ) (16d)
In which an asterisk * is a convolution operator, while ↓2,1
(↓1,2) is a sub-sampling operator along a row (or column)
and L0 = I(x) defines a genuine image. H and G are low pass
and band-pass h(k) and g(k)) filters on a horizontal or
vertical angle.
Ln is obtained from filtration of low-pass by h(k) and thus
regarded as an approximation image at scale n. Dni is
obtained from filtration of band-pass by g(k) on certain
angle and therefore contains detailed angle information at
scale n. It is then defined as a detailed image. A genuine
image I is represented by a group of sub-images at various
scales: {Ld, Dni}i=1,2,3; n=1..d which represents a compilation
of multi-scale sub-images with certain depth d from an
image I.
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
19
Types of main wavelet applied in this research include
Daubechies, Coiflets, and Symlets, and the character used is
normalized energy character of a sub-image Dni defined as
the below formula:
E ni =
1
∑ ( Dni (b j , bk )) 2
N j ,k
(17)
With a condition that N represents the total number of
wavelet coefficients of a sub-image Dni. Physically, this
wavelet energy character shows the spread of energy along
frequency axis towards scale and orientation.
D. Fractal Analysis Fractal geometry is a concept to explain geometrical
dimensions of objects that have scaling features/characters
or self-resemblance. This is frequently found on natural
object as well as human body. Different from Euclidean
geometry, fractal geometry is recognized for the value of
object dimension in fraction. In Euclidean geometry
concept, the smaller the scale within the measuring
instrument is used, the more precise the measurement
process of object dimension will be. In contrast, this
condition doesn’t apply to fractal objects. Therefore, it is
necessary to have fractal analysis methods. In this research,
two fractal analysis methods are used to obtain information
about bone textural characters: semi-variance method and
the method with Fourier transform approach [2].
1.) Semi variance
Bone texture analysis with semi-variance technique is
implemented on the basis of the analysis of neighboring
correlation among pixels in certain orientation. In this
technique, the value of semi-variance SV (θ, h) is calculated
from the sum of the subtractions of one pixel intensity and
its neighboring intensity on an image in distance h and
orientation θ. The orientation values commonly used in
most research are both 0° and 90° as shown below
SV (θ , h) =
1
2 (18)
∑(I (xi , y) − I (xi − h cosθ , y − h sinθ ))
2N N
According to the above formula, exponent Hurst is
determined by the observation of coefficient modulus of
Fourier transform from input signals and the observation of
dense curve gradient of spectral power on logs scale
towards signal frequency. Fractal analysis is done through
the observation of spectral power S(u,v) as
(21)
S 2 (u , v ) = F (u , v ) F * (u , v )
The value of spectral power obtained is subsequently
averaged on all angular orientations and this results in
S(|ω|). Signal dimension is determined from logarithm curve
slope S(|ω|) towards logarithm |ω|
(22)
D SA = 2 − slope
IV. EXPERIMENT
Experiment data is obtained as 37 x-ray film images
collected through taking images of 37 women with Singh
index of 2-5. The images are taken with analog x-ray
instrument, with output of film (radiograph). Each film is
then scanned to be digital data with scanner of 400 dpi
resolution and 8 bit intensity depth. Afterwards, digital data
achieved is processed in a set of the following stages: (1)
determination of RoI, (2) enhancement of image quality, (3)
image character extraction, and (4) classification. The result
of classification is eventually compared with the diagnosis
result carried out by a radiologist before.
A. RoI Determination
The process of RoI (Region of Interest) determination is
aimed at obtaining particular observation areas within
human body which are suitable for osteoporosis
interpretation. One of the areas commonly observed by
doctors to diagnose osteoporosis is Ward triangle. In this
research, the observation area is a square area of 256 × 256
pixels, including Ward triangle. Physically, this size
represents a square with length of 1.6256 cm each. RoI area
used in this research is illustrated in Figure 4.1.
In which N defines the size of an image influenced by the
number of pixels within it.
The calculation of fractal dimension value DSV(θ) from
the slope of logarithm curve SV(θ,h) toward logarithm h is
reflected by the following formula
DSV (θ ) =
4 − slope
2
(19)
2.) Fourier Transform
One of the most popular fractal signals is a signal fBm
(fractional Brownian motion). Signal fBm is largely
recognized because it is non-stationary, self-similar,
derivative-less, and represented by a scaling parameter
called Hurst exponent. Signal fBm is regarded having Hurst
exponent with the value H, in which the equation 0 < H < 1
applies, if the spectral density of the signal power complies
with the following formula [12]
Sx ~
1
f
2 H +1
(20)
Fig 4.1 llustration of RoI area used in the research
For easy use and interface purposes, RoI determination
is carried out in two stages:
• In stage one, users are required to push the central
point of RoI square.
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
20
Proceedings of CITEE, August 4, 2009
•
In stage two, users are obligated to decide the
angular point of a square, which is the point
directing to body axis (proximal direction).
Figure 4.2. shows illustration of RoI determination process
within the program created
Input and output of image quality enhancement processing
are illustrated in Figure 4.4. below:
Fig 4.4. Image Quality Enhancement Processing:
Input (left) and Output (Right)
C.
Fig 4.2. Illustration of RoI determination process
B. Image Quality Enhancement The second stage of digital data processing is image
quality enhancement. This process is implemented through
two important methods (median filtering and contrast
stretching) toward a RoI image of 256 × 256 pixels. Median
filtering is intended to reduce noise disturbance to the
image, while contrast stretching to standardize image
intensity. The former method is used with the support of
checking window of 5 × 5 pixels, and the latter one runs by
following the below equation:
I ′( x, y ) =
I ( x, y ) − µ (I ( x, y ))
× 95 + 160 (23)
max (I ( x, y )) − µ (I ( x, y ))
The above equation defines that I’ is pixel intensity that has
been processed and I initial intensity. Operators µ and max
refer to average and maximum operators. Number 160
represents expected average intensity for RoI which has
been processed; meanwhile, number 95 is the dynamic
intensity as a result of substraction of the value µ and the
highest intensity value within RoI. Both numbers are
achieved on the basis of a series of image quality
enhancement experiments towards 111 RoI areas used in the
research. Figure 4.3. illustrates the change of histogram
shape in the process of image quality enhancement
Fig 4.3 Change of intensity histogram at RoI area in contrast stretching,
before (left) and after (right)
ISSN: 2085-6350
Character Extraction Character extraction process is an integrated part of
trabekula textural analysis. This process is applied to an
image of which the quality has been enhanced. With the
four methods described previously, the research comes up
with 175 textural characters:
• Co-occurrence analysis: 6 features
•
Gabor filter: 20 characters, achieved by combining
4 orientations with 5 middle frequency values
(normalized)
•
Wavelet transform: 104 characters as a
combination of 2 coefficients (approximation and
detail), 13 kinds of the main wavelet, and 4
observation levels
•
Fractal analysis: 45 characters, obtained from a
combination of 2 methods (semi-variance and
fourier methods), each of which is enriched with
30 and 15 resolutions.
D. Classification Classification process is carried out with the help of
artificial neural system, exercised with the use of exercising
data in the form of 111 RoI images. Testing is done toward
740 testing sampels. Before classification process is
conducted, selection of certain characters significantly
correspondent with Singh index is done. By the selection,
there are 8 characters used in the research:
1. Level 3 approximation image energy in wavelet
transform with the use of the main wavelet Symlets
of orde 4.
2. Level 3 approximation image energy in wavelet
transform with the use of the main wavelet Symlets
of orde 5.
3. Combined fractal dimension in semi-variance
method with the experiment value of 50 pixels.
4. Vertical fractal dimension in semi-variance method
with the experiment value of 50 pixels.
5. Horizontal fractal dimension in Fourier analysis
method with the experiment value of 12 Fourier
coefficients.
6. Image energy in Gabor filter with the orientation of
135° and normalized middle frequency F = 2−4.5.
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
21
7.
Level 2 approximation image energy in wavelet
transform with the use of the main wavelet
Symlets of orde 3.
8. Inverse Different Moment (IDM) in co-occurrence
analysis.
The exercise of the artificial neural system is run using
propagation algorithm of Levenberg Marquardt with the
use of Matlab software. The structure of the artificial neural
system is composed with 3 layers. The first layer (input
layer) consists of 8 neurons; the second layer 12 neurons;
and the third layer (output layer) 6 neurons (equal to the
number of grades within Singh index).
V. RESULT Testing of 740 data samples (i.e., RoIs which have been
determined before as testing samples) results in textural
character classifications as shown in Table 5.1.
Table 5.1. Result of Singh Index Classification with the help of the
artificial neural system
Results of neural system diagnosis
by radiologists
Results of neural system diagnosis
In summary, the classification process ran quite well in
that 96,62 % of testing data (715) was classified properly;
3,11 % (23) with diagnosis error of one grade Singh index;
and 0,27% (2) with diagnosis error of 2 grade Singh index.
VI.
CONCLUSION This research succeeded in creating initial construction
and implementation of trabekula textural analysis method of
femur x-ray image, meaning that it was successful in
helping doctors do early osteoporosis detection. In this
research, a series of experiments and analyses were
conducted using various methods of image analysis in an
effort to identify trabekula textural characters in femur
proximal bone correlated with osteoporosis detection.
Image analysis methods were a combination of cooccurrence technique, Gabor filter, wavelet transform, and
fractal analysis. The implementation and testing of the
neural system resulted in system potentials that might help
early osteoporosis detection process now that 96,62% of
data samples were accuraely detected (similar to the result
of diagnosis by doctors/radiologists).
1
2
3
4
5
6
[1]
1
0
0
0
0
0
0
[2]
2
0
39
1
0
0
0
3
0
0
111
7
2
0
4
0
0
0
345
15
0
5
0
0
0
13
207
0
6
0
0
0
0
0
0
[3]
[4]
[5]
[6]
REFERENCES R.M. Haralick, K. Shanmugam, I. Dinstein, Textural Features for
Image Classification, IEEE Transactions on Systems, Man, and
Cybernetics, Vol.SMC-3, No.6, pp.610-621, 1973.
J.C. Lin, S. Grampp, T. Link, M. Kothari, D.C. Newitt, D. Felsenberg, S.
Majumdar, Fractal Analysis of Proximal Femur Radiographs:
Correlation with Biomechanical Properties and Bone Mineral
Density, Osteoporos Int, 9:516-524, 1999.
B. Sankaran, Osteoporosis, Mumbai: Novelty Printers, 2000.
---------, Modul Praktikum Analisis Tekstur, Imaging & Image
Processing Research Group, Institut Teknologi Bandung
T.R. Mengko, J.T. Pramudito, Texture Analysis of Proximal Femur
Radiographs for Osteoporosis Assessment, WSEAS Transaction on
Computers, Vol.3, No.1, 2004
S.W. Simson, M. Hidayat, Bone Architecture, Matrix Property and
Fracture Risks, Pertemuan Ilmiah Tahunan Perhimpunan
Osteoporosis Indonesia (PEROSI), 2007
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
22
Proceedings of CITEE, August 4, 2009
Optical Communication System on Microcell
Networks in Urban Area
Purnomo Sidi Priambodo, Harry Sudibyo and Gunawan Wibisono
Department of Electrical Engineering, Faculty of Engineering, Universitas Indonesia
Depok, Jawa Barat 16424, Indonesia
Email : [email protected]
HP : 0813-81153343
Abstract— The free-space optical communication (FSOC)
system is a new kind of technological engineering based on
the existing fiber-optic technology to obtain a breakthrough
technology which is wireless, flexible, less expensive and
large
bandwidth
capacity.
Free-space
optical
communications is a line-of-sight (LOS) technology that
transmits a modulated beam of visible or infrared light
through the atmosphere for broadband communications.
The entire components used in this optical free-space
communication system are taken and modified from the
existing fiber-optics system. Similar to fiber optical
communications, free-space optics uses a light emitting
diode (LED) or laser semiconductor diode source for data
transmission. Free-space optical communications offer data
rates comparable to fiber optical communications at a
fraction of the deployment cost of microwave or fiber-optics
links. The other advantages are flexible installation and
extremely narrow laser beam widths provide no limit to the
number of free-space optical links that may be installed in a
given location. Thus advantages are justified and suitable to
be implemented on microcell networks in urban areas.
The fundamental limitation of free-space optical
communications arises from the atmosphere through which
it propagates. Free-space optical communication systems
can be severely affected by fog and atmospheric turbulence.
The main design challenges in the free-space optical
communications are solving to overcome the effects of thus
fog, scintillation and beam wander.
In this paper, the authors elaborate optical free-space
propagation parameters such as attenuation, polarization
effect, Rayleigh scattering and possible fading. Further, the
mathematical model is introduced for each component and
path of the communication system. The model will be used
to design the optimal optical free-space communication
system implemented on microcell networks in urban area.
Keywords—free-space
optical communication, atmospheric propagation, atmospheric turbulence, spatial
diversity reception
I. INTRODUCTION
Free-space optical communications (FSOC) is a lineof-sight (LOS) technology where a modulated beam of
visible or infrared light is transmitted through the
atmosphere for broadband communications [1]. Similar to
fiber optical communications, free-space optical
ISSN: 2085-6350
communication uses a light emitting diode (LED) or laser
source for data transmission [2]. The beam is collimated and
transmitted through space or atmosphere rather than being
guided through a fiber-optics cable. Free-space optical
communications offer data rates comparable to fiber optical
communications at a fraction of the deployment cost of
microwave or fiber-optics links. The other advantages of
this technology are flexible installation and extremely
narrow laser beam widths provide no limit to the number of
free-space optical links that may be installed in a given
location. Thus advantages are justified and suitable to be
implemented on microcell networks in urban area.
At this time, this technology can be installed licensefree worldwide, can be installed in less than a day. This
line-of-sight technology approach uses invisible beams of
light to provide optical bandwidth connections [3]. It's
capable of sending up to 1.25 Gbps of data, voice, and video
communications simultaneously through the air — enabling
fiber-optic connectivity without requiring physical fiberoptic cable. This optical connectivity doesn't require
expensive fiber-optic cable or securing spectrum licenses
for radio frequency (RF) solutions. While fiber-optic
communications gained worldwide acceptance in the
telecommunications industry, FSOC is still considered
relatively new. This FSOC technology approach has a
number of advantages:
1.
2.
3.
4.
5.
Requires no RF spectrum license.
Easy to upgrade, and its open interfaces support
equipment from a variety of vendors.
It is immune to radio frequency interference or
saturation.
It can be deployed behind windows, eliminating the
need for costly antenna building such as for RF
technology.
No fiber-optic cables deployment required.
A. Applications:
FSOC technology can be applied for two different
implementations. The first are applications for space
communications between satellites, spacecrafts and deepspace between planets. The second are applications for
terrestrial communications, similar to point-to-point or lineof-sight in microwave communication technology. For
space communications, there are some facts that free-space
optical communications will enable space missions to return
10 to 100 times more data with 1% of the antenna area of
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
current state-of-the-art communications systems, while
utilizing less mass and power. Data can also be exchanged
between a more remote spacecraft and a station on /or near
Earth. For example, planetary probes can generate a lot of
image data, and a major challenge is to send large amount
of data back to Earth. Until recently, radio links operating
e.g. in the X band or Ka band were the only available
technology. The basic advantage of the optical technology
over radio links is that the much shorter wavelength allows
for a much more directional sending and receiving of
information. In technical terms, the antenna gain can be
much higher. This is particularly important for bridging
interplanetary distances. In the future FSOC will become
the fundamental communication technology for deep-space
communications.
The second are applications for terrestrial
communications and omni-wireless, such as point-to-point
communications between base-stations in cellular networks
and omni Wi-Fi Local Area Networks (LANs). When the
traffics increase, the cellular networks need to migrate
become microcellular networks where the cell areas is about
less than 1-km2. The number of cellular base-stations
(BTSs) increases in large number, the consequence is to
deploy the communication channels between BTSs. Fiberoptics deployments are not flexible and very expensive for
short distance between BTSs, while point-to-point
microwaves do not give a good large bandwidth. The
answer is FSOC, which gives us flexibility such as point-topoint microwaves, large bandwidth, easy deployment, not
heavy and inexpensive investment. Applications for
terrestrial are more challenging than applications for space
communications, because terrestrial applications facing the
effects of atmospheric turbulence.
The fundamental limitation of free-space optical
communications for terrestrial links arises from the
atmospheric medium through which it propagates. Freespace optical communication systems can be severely
affected by fog and atmospheric turbulence [4]. Fog is vapor
composed of water droplets, which are only a few hundred
microns in diameter but can modify light characteristics or
completely hinder the passage of light through a
combination of absorption, scattering, and reflection. This
can lead to a decrease in the power density of the
transmitted beam, decreasing the effective distance of a
free-space optical link. The atmospheric turbulence is
mostly due to wind and temperature gradient, causes the
index of refraction gradient by time and space. This
refractive index gradient deflects and diverge the light
beam, while the turbulence-wind will cause the light beam
wandering.
The items that limit the FSOC performance, then listed
in the following systematical parameters:
1) Absorpsion occurs when suspended water molecules in
the terrestrial atmosphere absorb photons. This causes a
decrease in the power density (attenuation) of the FSO
beam and directly affects the availability of a system.
Absorption occurs more readily at some wavelengths
23
2)
3)
4)
5)
than others. However, the use of appropriate power,
based on atmospheric conditions, and use of spatial
diversity (multiple beams within an FSO-based unit)
helps maintain the required level of network
availability.
Scattering is caused when the wavelength collides with
the scattering-particles. The physical size of the
scattering-particles determines the type of scattering.
When the particles are smaller than the wavelength, this
is known as Rayleigh scattering. When the particles are
of comparable size to the wavelength, this is known as
Mie scattering. When the particles are much larger than
the wavelength, this is known as non-selective
scattering. In scattering — unlike absorption — there is
no loss of energy, only a directional redistribution of
energy that may have significant reduction in beam
intensity for longer distances.
Polarization-rotation is caused by water molecules in
the terrestrial atmosphere act like dipole-molecules
with arbitrary orientation. Statistically it change the
polarization of the light beam.
Scintillation is the temporaland spatial variation in
light intensity caused by atmospheric turbulence. Such
turbulence is caused by wind and temperature gradients
that create pockets of air with rapidly varying densities
and, therefore, fast-changing indices of optical
reflection. These air pockets act like lenses with timevarying properties and can lead to sharp increases in the
bit-error-rates of free-space optical communication
systems, particularly in the presence of direct sunlight.
Beam-Wandering arises when turbulent wind current
(eddies) larger than the diameter of the transmitted
optical beam cause a slow, but significant,
displacement of the transmitted beam. Beam wander
may also be the result of seismic activity that causes a
relative displacement between the position of the
transmitting laser and the receiving photodetector.
Another concern in the implementation of FSOC is
safety, because the technology uses lasers for transmission.
The proper use and safety of lasers have been discussed
since FSO devices first appeared in laboratories more than
three decades ago. The two major concerns involve eye
exposure to light beams and high voltages within the light
systems and their power supplies. Strict international
standards have been set for safety and performance, comply
with the standards.
The main design challenges in FSOC are solving to
overcome the effects of thus absorption, scattering,
polarization rotation, scintillation and beam-wandering,
which are due to the existence of fog and atmospheric
turbulence. The first step prior to overcome the combination
of atmospheric disadvantage parameters is by creating
propagation model as explained in the next section.
B. Transmission Issues in Terrestrial Free Space Optical
Communications
Particularly for large transmission distances, it is
essential to direct the energy of the sender accurately in the
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
24
Proceedings of CITEE, August 4, 2009
form of a well-collimated coherence light beam. For long
distance (5-km) applications with large bandwidth, it is
suggested to use lasers that have narrower bandwidth and
long coherence-length. However, for applications those
require inexpensive investment with shorter distance and
smaller bandwidth, it is sufficient to use collimated LED.
The collimated light beam is required in order to limit the
loss of power between the sender and the receiver due to
beam expansion. However, the one needs large beam radius
with high-quality optical front-wave to keep the receiver in
line with the transmitter when beam-wandering happens.
The one should use a diffraction-limited light source and a
large high-quality optical telescope for collimating a large
beam radius. Due to the short wavelength of light, the beam
divergence of an optical transmitter can be much smaller
than that of a radio or microwave source of similar size.
Using a frequently used term in the area of radio
transmission, the antenna gain can be much higher for
optical transmitters over 100 dB even for moderate
telescope diameters of e.g. 25Ԝcm.
Fig. 1: Setup for a Simplex FSOC system. Although the
transmitter signal is approximately collimated, part of the
transmitted power may miss the detector.
It is also advantageous to have a large high-quality
optical telescope on the side of the receiver. It is essential
not only to collect as much of the sender's power as
possible, but also to minimize disturbing influences, e.g.
from background light, which introduces noise and thus
reduces the data transmission capacity. Both high sensitivity
and high directionality can be achieved by using a large
high-quality telescope at the receiver end. High
directionality also requires high precision in the alignment
of the sender and receiver.
An important issue is the power budget of a free-space
link, including the transmitter's power and all power losses.
The remaining power at the receiver largely determines the
possible data transmission rate, even though this is also
influenced by the modulation format, the acceptable bit
error rate, and various noise sources, in particular laser
noise, pre-amplifier noise, excess noise in the receiver (e.g.
an avalanche photodiode), and background light. The latter
can often be efficiently suppressed with additional narrowband optical filter and high-quality telescope, since the
optical bandwidth of the signal is fairly limited, whereas
background light is usually very broadband.
Severe challenges can arise from the effects of
atmospheric disturbances such as clouds, dust and fog,
which can cause not only strong signal attenuation but also
inter-symbol interference. To solve this problem, it is
suggested to use sophisticated techniques of forward error
correction embedded in the datalink layer, which allow for
reliable high-capacity optical links even through thick
clouds.
II.
SPATIAL DIVERSITY RECEPTION MODEL TO
IMPROVE PERFORMANCE
In this section, the authors will elaborate a statistical
model of spatial diversity reception to overcome the effects
of atmospheric turbulence. Spatial diversity reception,
which has been introduced for RF applications, is a
potential method to overcome the degradation due to
atmospheric turbulence [5-9]. Application of spatial diversity
reception in FSOC has been introduced in some researches
[7-9]
. The performance of spatial diversity reception on
FSOC has been studied and proposed [8], Assuming that
atmospheric-turbulence fading is uncorrelated at each of the
optical receivers. In order for this assumption reasonable,
the spacing between receivers should exceed the fading
correlation length in the plane of the receivers. It may be
hard to satisfy this assumption in practical situation, for any
reasons. Available space may not permit enough receivers
spacing. In power-limited links, which employ collimated
beam with large diameter, the receiver spacing required
obtaining uncorrelated fading, may exceed the beam
diameter.
A. Maximum-Likelihood (ML) Diversity Detection on
Turbulence Channels
In the case of spatial diversity reception with nreceivers, the received signal is presented by n-component
vector. Taking account of correlation between the receivers,
bit-error probability equation for maximum-likelihood in [ 4]
is modified as follows:
r
⎡ n r2 ⎤ n
1
P r | Off = exp ⎢ −∑ i ⎥ ∏
⎣ i =1 Ni ⎦ 1 π Ni
(
)
(
⎡ n r − η I e 2 X i −2 E[ X i ]
r
uur
i
0
P r | Off = ∫ f uuXr X • exp ⎢ − ∑
⎢ i =1
uur
Ni
X
⎣⎢
(
)
( )
) ⎤⎥ •
(1)
2
⎥
⎦⎥
n
∏
1
dX i
π Ni
(2)
Where Ni / 2 is noise covariance of the i-th receiver and
exp{− 12 ⎣⎡( X 1 − E [ X 1 ])
uur
f uuXr X =
( )
⎡ ( X 1 − E [ X 1 ]) ⎤
⎢
⎥
− E [ X n ]) ⎦⎤} C X−1 ⎢
⎥
⎢ X − E[X ] ⎥
n )⎥
⎢⎣( n
⎦
n/2
( 2π ) | C X |1/ 2
(X
n
(3)
CX is the covariance matrix of the log-amplitude in the nreceivers. The likelihood function
ISSN: 2085-6350
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
r
P r | On
r
Λ r = r
P r | Off
(
(
()
= ∫ f uuXr
uur
X
25
)
)
where
(4)
(
⎡ n r − η I e2 X i − 2 E[ X i ]
uur
0
i
X • exp ⎢ −∑
⎢ i =1
Ni
⎢⎣
( )
)
2
− ri ⎤ uur
⎥dX
⎥
⎥⎦
2
uur
⎛ T ⎞
P Bit Error X = P ( Off ) Q ⎜ th ⎟ + P ( On )
⎝ 2N ⎠
⎛ η I 0 e 2 X1 − 2 E[ X1 ] + e 2 X 2 − 2 E[ X 2 ] − Tth
Q⎜
⎜
2N
⎝
(
)
(
)
(10)
⎞
⎟
⎟
⎠
B. Numerical Simulations for Multi Receivers
The ML detector uses the decision rule
r > On .
Λ r
1
< Off
()
Since the
log-amplitude follows a joint log-normal distribution,
calculation of the likelihood function in (4) involve multidimensional integration. It is emphasized that this decision
rule has been derived under the assumption that the
receiving side knows the fading correlation but not the
instantaneous fading state.
The bit-error probability of the ML receiver is given by
Pb = P(Off) . P(Bit Error| Off) + P(On) . P(Bit Error|On) (5)
where P(Bit Error| Off) and P(Bit Error| On) denote the biterror probabilities when the transmitted bit is off and on,
respectively. Without considering inter-symbol interference,
which can be ignored when the bit rate is not high and
multipath effects are not pronounced, we have:
P ( Bit error Off ) =
(
)
∫
r
r
p r Off d r
∫
r
r
p r On d r
r
Λ r >1
()
In this section, we intend to present numerical
simulations of the performance of spatial-diversity
reception/detection with multi-receivers and dual-receivers
as well. However, up to this Conference submission
deadline our research group has not finished the numerical
simulation yet. The structure of multi-receiver link that is
represented by dual-receiver is illustrated by Fig.2. As
described above, the ML receiver (Fig.2a) has full
knowledge of the turbulence-correlated fading matrix CX,
while the EGC receiver (Fig.2b) has knowledge only the
marginal distribution of fading at the receivers. It is
assumed that E[I1] = E[I2] = E[I] and N1 = N2 = N and
define the electrical signal-to-noise ratio SNR = (ηE[I]2)/N.
(6)
and
P ( Bit error On ) =
r
Λ r <1
()
(
)
(7)
To evaluate the optimal ML diversity detection scheme, we
compare it with the conventional equal-gain combining
(EGC) scheme [ ]. In the EGC scheme, it is assumed that
the receiving side has knowledge of the marginal
distribution of the channel fading, in every receiver, but has
no knowledge of the fading correlation or the instantaneous
fading state. For each single individual receiver output, it is
able to find an optimum threshold τi. Then the EGC detector
adds together the n-receivers output with equal gains and
compares the sum to the threshold
n
Tth = ∑τ i
(8)
i =1
The error probability of EGC is
uur
uur uur
Pb = ∫ f uuXr X P Bit Error | X d X
( ) (
uur
X
III.
)
(9)
CONCLUSION
Free space optical communications through atmosphere
is under intensive researches, where various methods have
been
proposed
to
overcome
turbulence-effects
communication signal fading. In this paper we introduce
Fig. 2: Dual-receiver model on atmospheric turbulence channels
with correlated turbulence fading. (a) Maximum-likelihood
detection. (b) Equal-Gain combining with threshold detection.
To obtain simulation results, it is assumed that E[X]= 0
and σX = 0.1, varying the normalized correlation ρd from 0
to 0.9. It is also compared to set up value δX = 0.25. We
expect that turbulence correlated fading will cause a larger
degradation of bit error probability when the standard
deviation noise δX is larger. Hipothetically, with tworeceivers, ML detection achieves a better performance than
EGC for a given SNR. The advantage of ML over EGC
method is more pronounced when the correlation ρd
between the two receivers is high. It is more reasonable
when the SNR is high, then the errors are caused mainly by
turbulence-correlated fading, as opposite to the noise.
spatial diversity reception model for free space optical
communications. This model can be use to analyze and later
on to overcome the effects due to turbulence-correlated logamplitude fluctuations.
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
26
Proceedings of CITEE, August 4, 2009
ACKNOWLEDGMENT
This research is supported by Universitas Indonesia
Research Grant called RUUI-2009, Nb.DRPM/RUUI
Unggulan/2009/I/3342 .
[5]
G. L. Stuber, Principles ofMobile Communication. New York: Kluwer
Academic, 1996.
[6]
J. G. Proakis, Digital Communication, 3rd ed. New York:
McGraw-Hill, 1995.
[7]
M. Srinivasan and V. Vilnrotter, “Avalanche photodiode arrays for
optical communication receivers,”, NASA TMO Progress Rep. 42144, 2001.
[8]
M. M. Ibrahim and A. M. Ibrahim, “Performance analysis of optical
receivers with space diversity reception,” Proc. IEE—Commun., vol.
143, no. 6, pp. 369–372, December 1996.
[9]
X. Zhu and J. M. Kahn, “Maximum-likelihood spatial-diversity
reception on correlated turbulent free-space optical channels,” presented
at the IEEE Conf. on Global Commun., San Francisco, CA, Nov.–Dec.
27–1, 2000.
REFERENCES
[1]
V. W. S. Chan, “Optical Space Communications”, IEEE J. Sel. Top.
Quantum Electron. 6 (6), 959 (2000)
[2]
O. Caplan, “Laser communication transmitter and receiver design”, J.
Opt. Fiber Commun. Rep. 4, 225 (2007)
[3]
K. F. Büchter et al., “All-optical Ti:PPLN wavelength conversion
modules for free-space optical transmission links in the midinfrared”, Opt. Lett. 34 (4), 470 (2009)
[4]
Xiaoming Zhu and Joseph M. Khan, “Free Space Optical
Communication Through Atmospheric Turbulence Channels,” IEEE
Transactions on Communications, Vol. 50, No. 8, August 2002
ISSN: 2085-6350
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
27
A Measure of Vulnerability For Communication Networks:
Component Order Edge Connectivity
A. Suhartomo, Ph.D.
Department of Electrical Engineering
President University
[email protected],
[email protected]
Abstract: Graph theoretical techniques provide a
convenient tool for the investigation of a communication
network’s vulnerability to failure. When there are point to
point connections, a communication network can be
modeled by a graph. A measure of the vulnerability of a
network to link failure is the so-called edge connectivity of
the graph. We introduce a new vulnerability parameter,
generalization of edge connectivity, called “Component
Order Edge Connectivity”. It is the smallest number of
edges that must be removed in order to ensure that all
resulting components have fewer nodes than some given
threshold value. We then derive formulas for λc( k ) for stars,
paths, cycles, and complete graphs; but no formula has been
found for an arbitrary graph G.
Keywords: edge connectivity; minimum degree; kcomponent edge connectivity; edge-failure set, edge-failure
state
1.
Introduction
Networks such as ring fiber optic, satellite
communication, terrestrial microwave, and social all can be
modeled using graph-theoretic construct.
Network
vulnerability
is
an
important
consideration in network design. The utilization of
communication networks has grown tremendously in the
last decade, for activities such as transmitting voice, data,
and images around the world. With the widespread
dependence upon such networks, it becomes important to
find topologies that yield a high level of reliability and a
low level of vulnerability to disruption.
It is desirable to consider quantitative measures of
a network’s vulnerability. To obtain such measures we
model the network by a graph in which the station terminals
are represented by the nodes of the graph and the links are
represented by the edges.
The problem, in which edges are subject to failure but nodes
are not will be investigated here as a new vulnerability
parameter of networks called “k-component order edge
connectivity” or “component order edge connectivity”.
This is a departure from the classical edge failure model,
but we make use of edge connectivity and are concerned
with how these parameters interrelate.
2.
Network vulnerability in practice
Graph theoretic techniques provide a convenient
tool for the investigation of the vulnerability of a
communication network to damage from natural or enemy –
induced damage. Thus, a communication system can often
be represented as interconnection of stations and links. The
assumption is that the system is subject to natural failure or
enemy attack aimed at isolating stations from each other.
The problems of determining the vulnerability and
designing communication network which are invulnerability
to enemy attack are of paramount importance. As yet,
neither the analysis nor synthesis problems have been
completed proved, although a number of practical results
have been discovered. One difficulty immediately
encountered in vulnerability studies is lack of a completely
appropriate vulnerability criterion.
In this paper, we attempt to address this deficiency
by introducing a new parameter: the minimum number of
edges that must be deleted in order to isolate a given
number of nodes, called component order edge
connectivity.
3.
Graph Theory
Background
and
Network
Vulnerability
This section introduces graph theoretic concepts
which are fundamental to understanding network
vulnerability, and surveys the relevant existing work in the
field.
A graph G = (V,E), or G, consists of a finite nonempty set of nodes V and a set E of two element subsets of
V. If {u , v} ∈ E , we say that {u , v} is incident at the
nodes u and v, and that u and v are adjacent. If
| V |= n, | E |= e , G is referred to as an (n,e) graph; n is
the order of G and e is size of G.
A specific example of a graph G is given by
V = {v0 , v1 , v 2 , v3 , v 4 , v5 , v6 , v7 } , and
E = {{v0 , v1 }, {v0 , v 2 }, {v0 , v3 }, {v1 , v5 },
{v 2 , v5 }, {v 2 , v 4 }, {v3 , v 4 }, {v 4 , v5 }, {v 4 , v6 },
{v5 , v7 }, {v 6 , v7 }}.
If it is convenient to do so, we usually represent
the graph pictorially, e.g., Figure 1 depicts the graph given
in the exam
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
28
Proceedings of CITEE, August 4, 2009
v1
v0
v2
v5
v7
neighborhood of a node v, denoted by N(v), is the set of
nodes which are adjacent to v. The degree of a node is thus
the cardinality of its neighborhood, i.e., deg v = |N(v)|. The
minimum degree among the nodes of G is denoted δ (G) =
min{deg v| v ∈ V} and the maximum degree is denoted
∆ (G) = max{deg v| v ∈ V}.
Since each edge has two end-nodes, the sum of the
degrees is exactly twice the number of edges:
n
∑ deg v
v3
v4
v6
(1)
⎢ 2e(G ) ⎥
⎣ n ⎥⎦
δ (G ) ≤ ⎢
(2)
⎡ 2e(G ) ⎤
.
∆ (G ) ≥ ⎢
⎢ n ⎥⎥
(3)
and
•
Connectedness
A desirable property for networks such as
communication networks is that each station be able to send
messages to all other stations, if not directly then by
relaying through other stations. In terms of the graph that
models the network, the desired property is that the graph is
connected. Therefore, we introduce some basic notions
pertinent to connectedness.
A walk in a graph G is an alternating sequence of
nodes and edges,
v0 , x1 , v1 , x 2 , L , v n −1 , x n , v n , where xi = {vi −1 , vi } for
i = 1, L , n.
It is closed if v0 = v n , and is open otherwise. Since G has
no multiple edges or loops, it suffices to suppress the edges
and just list the nodes in order of appearance.
A path in a graph is walk in which no node is
repeated. If a walk is closed, then it is a cycle provided
v0 , v1 ,L, v n −1 are distinct and n ≥ 3 .
A graph G is said to be connected if for all pairs of
distinct nodes u and v there exists a walk (equivalently a
path) joining u and v. A disconnecting set of a graph is a set
of edges which renders the graph disconnected upon
removal.
In the labeled graph G of Figure 2.1.,
v0,v3,v4,v2,v4,v5,v7 is a walk, which is not a path, v0,v2,v4,v5,v7
is a path, and v2,v4,v5,v2 is a cycle. The graph G is connected
and {{v5 , v 7 }, {v 4 , v 6 }} is a disconnecting set.
•
Subgraphs
A graph H = (V(H),E(H)) is a subgraph of a graph
G, denoted H ⊆ G, if
V(H) ⊆ V(G) and E(H) ⊆ E(G). If V(H) = V(G), then H is
called a spanning subgraph. Therefore, a spanning
subgraph may be viewed as a graph obtained by the removal
of a set of edges. Observe that G is connected if and only if
G has one component.
Degrees
The degree of a node v in a graph G, denoted by
deg v, is the number of edges incident with v. In a (n,e)
graph, 0 ≤ deg v ≤ n-1 for every node v. The
ISSN: 2085-6350
= 2e(G ) .
Hence
Figure 1. A pictorial representation of a graph.
•
i
1
4.
Classical Network Vulnerability
The parameters (node) connectivity and edge
connectivity are used to measure the “vulnerability” of a
network (graph) to disconnection upon failure of nodes or
edges, respectively. Specifically, the (node) connectivity
κ (G ) is the minimum number of nodes required to be
removed so that the surviving subgraph is disconnected or
trivial (i.e., a single node). The edge connectivity λ (G ) is
the minimum number of edges required to be removed so
that the surviving graph is disconnected. He also proved that
⎢ 2e ⎥
⎣ n ⎥⎦
κ (G ) ≤ λ (G ) ≤ δ (G ) ≤ ⎢
(4)
Figure 2 provides an example of this result.
Figure 2. A graph with
κ (G ) = 1, λ (G ) = 2,
⎢ 2e ⎥
δ (G ) = 3, ⎢ ⎥ = 3
⎣n⎦
In many networks, disconnection may not
guarantee that the network can no longer perform the
function that it was designed for. As long as there is at least
one connected piece which is large enough, the network
may still be considered operational. Hence there are
inadequacies inherent in using connectivity or edge
connectivity as a measure of vulnerability. A concrete
example of this deficiency is show in Figure 3. The graph G
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
29
of order 1,000,001 has κ (G ) = λ (G ) = 1 . However the
subgraphs G – v and G – {u,v} have components of order
999,999 and 1,000,000, respectively.
component order edge connectivity of G, denoted by
λ(ck ) (G ) or simply λ(kc ) , is defined to
λ(ck ) (G ) = min{| F |: F ⊆ E , F is k - component
be
edge-failure set}, i.e., all component of G – F have order
≤ k −1.
u
v
K 10 6
Definition 3. A set of edges F of graph G is
Figure 3. A graph with
λ (G ) = 1
It is reasonable to consider a model in which it is
not necessary that the surviving subgraph is connected so
long as it contains a component of some predetermined
order. Component order edge connectivity is the minimum
number of edges required to be removed so that the
surviving subgraph contains no component or order at least
some prescribed threshold.
5.
Component Order Edge Connectivity
•
The Traditional Edge-Failure Model
In the traditional edge-failure model it is assumed
that nodes are perfectly reliable but edges may fail. When a
set of edges F fail we refer to F as an edge-failure set and
the surviving subgraph G – F as an edge-failure state if G –
F is disconnected.
Definition 1. The edge connectivity of G, denoted by
λ (G ) or simply λ , is defined to be λ (G ) = min{
| F |: F ⊆ E , F is an edge-failure set }, i.e., G – F
disconnected.
One drawback of the traditional edge-failure model
is that the graph G – F is an edge-failure state if it is
disconnected and no consideration is given to whether or
not there exists a “large” component which in itself may
be viable. In this paper we introduced a new edge-failure
model, the k-component edge-failure model.
•
The New Edge-Failure Model
It is reasonable to consider a model in which it is not
necessary that the surviving edges form a connected
subgraph as long as they form a subgraph with a
component of some predetermined order. Thus we
introduce a new edge-failure model, the
k-component order edge-failure model. In this model,
when a set of edges F fail we refer to F as a kcomponent edge-failure set and the surviving subgraph
G – F as a
k-component edge-failure state if G - F contains no
component of order at least k, where k
is a
predetermined threshold value.
λ(kc ) -edge set
if and only if it is a k-component order edge-failure set and
| F |= λ(ck ) .
Next we compute
λ(ck ) (G ) for specific type of
graphs. The first type of graph we consider is the star,
K 1,n −1 . Deletion of any set of m edges results in a subgraph
consisting of m+1 components, one isomorphic to K 1, n −m −1
and the remaining components isolated nodes. Therefore a
k-component edge-failure state exists if the component
K 1, n −m −1 contains at most k-1 nodes. Thus n − m ≤ k − 1
or n − k + 1 ≤ m . Since component order edge
connectivity is the minimum number of edges whose
removal results in a k-component edge-failure state we
obtain the following result:
Theorem 1: Given
2 ≤ k ≤ n, λ(ck ) ( K1,n−1 ) = n − k + 1
The next type of graph we consider is the path on
n nodes, Pn . Starting at a pendant edge label the edges
consecutively from 1 to n – 1. Let F be the set of edges
whose label is divisible by k – 1. The deletion of the edges
in F creates |F| + 1 components, the first |F| having order
exactly k-1 and the last order at most k-1 (but at least 1). We
also note that if fewer edges are removed then there will be
fewer components and thus by the Pigeonhole Principle,
there must be at least one component of order at least k.
⎢ n − 1⎥
we have:
Since |F| = ⎢
⎣ k − 1⎥⎦
Theorem 2: Given 2 ≤ k ≤ n,
⎢ n − 1⎥
.
⎣ k − 1⎥⎦
λ(ck ) ( Pn ) = ⎢
Theorem 3. Let Tn be a tree on n nodes. Then for any
value k, 2 ≤ k ≤ n,
⎢ n − 1⎥
(k )
(k )
(k )
⎢⎣ k − 1⎥⎦ = λc ( Pn ) ≤ λc (Tn ) ≤ λc ( K 1, n −1 ) = n − k + 1
The next type of graph considered is the cycle on n
nodes, C n . The formula for
similar manner to that of
λ(ck ) (C n ) can be derived in a
λ(ck ) ( Pn ) . When one edge of the
Definition 2. Let 2 ≤ k ≤ n be a predetermined threshold
value. The k-component order edge-connectivity or
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
30
Proceedings of CITEE, August 4, 2009
cycle is removed, it becomes the path with n nodes, Pn .
and
since
λ(ck ) (C n ) = λ(ck ) ( Pn ) + 1 ,
⎡ n ⎤
⎢ n − 1⎥
⎢⎣ k − 1⎥⎦ + 1 = ⎢⎢ k − 1⎥⎥ we have the following result:
Thus
Theorem 4. Given 2 ≤ k ≤ n,
⎡ n ⎤
.
⎢ k − 1⎥⎥
λ(ck ) (C n ) = ⎢
The last type of graph considered in this section is
the complete graph on n nodes, K n . Let F ⊆ E ( K n ) be
a
λ(kc ) -edge set set. We can compute |F| by calculating the
maximum number of edges that can remain in the kcomponent edge-failure state K n − F . It is easy to see
that any edge in F must have its endpoints in two different
components of K n − F ; thus each component of
K n − F must itself be complete. Thus we have the
following:
2 ≤ k ≤ n,
⎛ n ⎞ ⎢ n ⎥⎛ k − 1⎞ ⎛ r ⎞
λ(ck ) ( K n ) = ⎜⎜ ⎟⎟ − ⎢
⎥⎦⎜⎜ 2 ⎟⎟ − ⎜⎜ 2 ⎟⎟ ,
2
k
−
1
⎣
⎝ ⎠
⎝
⎠ ⎝ ⎠
n
⎢
⎥
where n = ⎢
(k − 1) + r , 0 ≤ r ≤ k − 2 .
⎣ k − 1⎥⎦
Theorem 5. Given
References
[1] Gross, D., Boesch, F., Kazmierczak, L., Suffel, C.,
Suhartomo, A.: “Component Order Edge Connectivity
– An Introduction”, Congressus Numerantium 178
(2006), pp. 7-14.
[2] Gross, D., Boesch, F., Kazmierczak, L., Suffel, C.,
Suhartomo, A.: “Bounds For Component Order Edge
Connectivity”, Congressus Numerantium 185 (2007),
pp. 159-171.
ISSN: 2085-6350
[3] Suhartomo, A.: “Component Order Edge Connectivity:
A Vulnerability Parameter
For Communication
Networks”, Ph.D Dissertation, Stevens Institute of
Technology, Hoboken – NJ, USA, 2007
[4] Frank Boesch, Daniel Gross, L. William Kazmierczak,
John T. Saccoman, Charles Suffel, A. Suhartomo.: “A
Generalization of A Theorem of Chartrand”, Network
– 2009-DOI 10.1002/net
Biography:
In 1975 – 1977, Antonius Suhartomo received training
courses in Radio Transmission of Microwave Terrestrial
and Satellite from PT. Telekomunikasi Indonesia
(TELKOM), and from 1977 – 2004 Antonius was an
employee of PT. TELKOM. He received a B.S in Physics
from University of North Sumatera, Medan – Indonesia in
1984, followed by a Master (M.Eng.Sc.) in OpticalElectronic and Laser Applications from University of
Indonesia, Jakarta – Indonesia in 1993. He also received a
Master (M.M.) in Management from Graduate School of
Management, Institute Management Telkom, Bandung –
Indonesia in 1994. At PT. TELKOM from 1977 – 2004, he
was assigned in many departments under TELKOM
business such as in Engineering Division, Training
Division, R & D, Institute of Technology Telkom, and
Maintenance Service Center Division before he took his
Ph.D. He received his Ph.D. in Electrical and Computer
Engineering Department majoring in Electrical Engineering
from Stevens Institute of Technology, Hoboken – NJ, USA
in 2007. He is currently in-charge in several positions such
as Secretary Department of Electrical Engineering, Head of
Research and Development Center, and Director Office of
Residence Life at President University Kota Jababeka –
Cikarang, Bekasi – Indonesia starting from year 2008..
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
31
The Speech Coder at 4 kbps using Segment
between Consecutive Peaks based on the
Sinusoidal Model
Florentinus Budi Setiawan
Electrical Engineering, Soegijapranata Catholic University,
Semarang 50234, Indonesia
[email protected]
Abstract— Communications equipment needs to work at
low bit rate, that can reduce the need of transmission
bandwidth. The proposed speech coder generates coded signal
at the rate of 4 kbps with low complexity. Thus, the
transmission channel can be used for great number of
communications connection. The low bit rate speech coder can
be realized by using segmental sinusoidal model. By using this
model, signal parameters are used to generate the synthetic
speech signal.. In this paper, we proposed a sinusoidal model
by using the peak of signal amplitude, called as segmental
sinusoidal model. The proposed coder is combined with
waveform interpolation model and codebook, thus it can work
at low bit rate. The coder has MOS score of 3.8 (out of 5). It
means that the perception quality is fairly good. The output
data rate of encoder is 4 kbps or bellow, with complexity less
than 10 MIPS(Million Instruction per Second).
Keywords—analysis,
frequency,
period, segmental, sinusoidal, synthesis
I.
interpolation,
peak,
INTRODUCTION
The number of transmission channel for communication
become limited due to increase of the communication
channel usage. Limited channel capacity had endorsed all
side more efficient. Speech signal data rate can be developed
by using encoder to obtain the simpler information. The
research aim is obtained the low rate speech signal encoder
with high quality.
The research aim is obtaining coding algorithm on the
rate of 4 kbps with high perception quality, on Mean
Opinion Score 3,5 - 4. Research consists of encoder and
decoder development at conversation frequency between 300
Hz and 3400 Hz. Coder system can be applied for speech
signal storage and it can be implemented at digital signal
processor system. On the other hand, coder can be applied
into communication system using digital signal
controller.The main contribution of the research is obtaining
the method for decreasing speech signal data rate at 4 kbps.
The other contribution is saving the storage media capacity.
The number of saving is 93% compared to G.711 speech
coder. The coder can be implemented into speech
communication hardware like cellular phone, cordless
phone, VoIP, multimedia system, secret communication and
many other communication system.
The speech coder system consists of encoder and
decoder. The encoder function is encoding the speech signal
at rate 4 kbps and the decoder function is converting the
coded signal into estimation speech signal approximate with
the original speech signal. Main part of encoder is
compression algorithm with segmental sinusoidal method for
speech signal analysis and waveform interpolation to reduce
data rate. Differentiation of the voiced signal and the
unvoiced signal is useful for compression method choosing.
The voiced, the unvoiced, the pitch, the segment, the
decimation, and the formant parameters are coded into
codebook. Decoder consists of parameters synthesis block
and quality enhancing block. The coded signal is synthesised
by using sinusoidal signal generator based on transmitted
parameters. Code-vectors are generated from the transmitted
codebooks index. Reconstructed speech signal enhancement
is developed as a four parallel of band-pass filters for
increasing the formant signal and decreasing the valleys
between formants.
The research consists of speech signal algorithm
development at the rate of 4 kbps, coder software simulation
and hardware realization at digital signal processor. The
simulation is developed by using the Borland C++ software
at 4.5 version. Software simulation consists of speech signal
acqusition, speech signal coder, speech signal reconstruction,
speech signal reproduction as results of the signal
reconstruction and performance tests. The hardware
implementation is applied at digital signal processor starter
kit (DSK) TMS320VC5416 from Texas Instrument by using
Code Composed Studio software. The simulation results are
implemented into DSK TMS320VC5416 hardware.
Based on the development and experiment results, I
conclude that speech signal coder at 4 kbps have realized by
using segmental signal model, waveform interpolation and
codebook usage. Near toll perception quality have realized
by using the voiced and the unvoiced differentiation,
harmonic signal of the sinusoidal model and postfilter
implementation. Coder system have applied for speech
signal storage and it have been realized by using a hardware
based on digital signal processor. The number of the
memory needed for encoder and decoder routines saving is
less than 16 KB, so that it is realizable for digital signal
processor system implementation. The coder system also
have been applied for communication system and it can be
realized by using the custom integrated circuit design. The
number of instruction is less than 6 million per second (6
MIPS), so that it can be implemented into the recent realtime
system that work at speed of more than 160 MIPS.
Suggestion for the futher research is coder implementation at
the digital signal controller, that consists of DSP, memory,
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
32
Proceedings of CITEE, August 4, 2009
ADC and DAC. So, it can be used as encoder and decoder at
application scale.
II.
SINUSOIDAL MODEL
Sinusoidal model can be developed to obtain less
parameters, so that the signal data rate can be reduced. This
model is called as segmental sinusoidal model. By using this
model, there are two harmonic signal to estimate the original
signal between two consecutive peaks, (maximum to
minimum or minimum to maximum). Peak means minimum
peak or maximum peak on the frame. Therefore, one
segment means part of signal between maximum peak and
consecutive minimum peak or part of signal between
minimum peak and consecutive maximum peak.
Time distance between i-th maximum peak and
consecutive minimum peak called as period information, and
denoted by pd(i). Maximum peak or minimum peak called as
peak information, and denoted by pk(i). Peak information is
obtained by detecting the maximum peaks and minimum
peaks over the frame observed. Period information is
obtained by counting the time distance between the
consecutive peaks. The proposed method is a process on
time domain. On the extreme waveform coding, signal is
fetched on its peaks [1]. On the one frame with length of N,
there are M maximum peaks and L minimum peaks. On this
frame, there are large number sinusoidal signal components.
It can be written as :
K
s(n) = ∑ ak cos(ωk (n)n + φk (n))
k =0
(1)
for : 0 < k < K-1, n = 0,1,2 … N-1 , K<N
The first and second coefficients (k=0 and k=1) are used
as components to reconstruct the estimated signal from one
maximum peak until the consecutive minimum peak or from
one minimum peak until the consecutive maximum peak.
This is the equation for estimated signal
s( n) = a0 + a1 cos(ω1 ( n) + φ1 ( n))
(2)
The estimated signal over the frame is a train of the spv
and svp for i=0 until i=I-1. based on the previous
explanation, for pk (0) > pk(1), the reconstructed signal using
the segmental sinusoidal model can be written as :
sr (i, n) =
⎞
pk (i ) + pk (i + 1) pk (i ) − pk (i + 1) ⎛ π .( n − nk (i ))
+
+ i.π ⎟⎟
cos⎜⎜
2
2
p d (i )
⎝
⎠
(3)
for i = 0,1,2 …(I-1)
If pk(0) < pk(1), the reconstructed signal using the
segmental sinusoidal model can be written as :
s r (i , n ) =
⎛ π .( n − nk (i ))
⎞
p k (i ) + p k (i + 1) p k (i ) − p k (i + 1)
+
+ (i + 1).π ⎟⎟
cos⎜⎜
2
2
p d (i )
⎝
⎠
for i = 0,1,2 …(I-1)
ISSN: 2085-6350
(4)
III.
SPEECH CODER AT 4 KBPS
In this paper, a speech signal encoder at 4 kbps and
bellow has been designed using several blocks and
algorithms. Detail of the encoder is shown in fig. 6. The
encoder contains existing signal detector, windowing
process, and pitch detector. The next blocks are voiced and
unvoiced classificator, sinusoidal based coder, and formant
coder.There are some operation mode of the encoder system
depends on the kind of signal to obtain the high performance
of coding system [2-7]. There are two different operation
modes: silent operation mode and signal operation mode.
The signal operation mode consists of vibrating mode
operation and non-vibrating mode operation. Input signal is
speech signal in 16-bit PCM format at 8 kHz frequency
sampling. The first block is signal buffer with 30 ms length.
The next block is existing signal detector. Then the 30 ms
signal will be detected its pitch period width. Based on pitch
period information, signal would be classified into vibrating
and non-vibrating signal. If it is less than 160 samples, the
signal in buffer is called as vibrating signal. Then, if it is
more than 160 samples, it is called as non-vibrating signal.
The next process is depend on the kind of signal. For
vibrating signal (voiced), characteristic signal [8][9] have to
be held. One pitch period of signal is quantized using
segmental sinusoidal model. The formant information for
each pitch period is kept to obtain the variation information
changing for 30 ms. The next block is codebook index
searching based on periods, peaks, and formants. All of the
coded parameter are sent to the decoder with rate 4 kbps or
less than 4 kbps, depends on the kind of the speech signal.
The speech input signal exist are detected by using
existing signal detector. The signals are buffered with length
of 30 ms. The detector identify the input speech signals
whether there are signals exist or there are no signals exist. If
there are no signals exist, they are called as silence. A sign is
transmitted into decoder to inform this condition, so that the
decoder is not process the signal during 30 ms. But if there
are signals exist, the encoding process is continued with
pitch detecting process.
Pitch is the useful parameter in encoding process. Based
on the pitch value, we would identify the signals whether
voiced or unvoiced. Human speech pitch period of voiced
part is vary from 2.5 ms until 20 ms, depend on the gender
and the age. Men tend to have the longer pitch period than
women and children [10]. The pitch period is detected by
using autocorrelation process. The first step, the buffered
signal is detected on its peak. Based on the peak value, it can
be found the threshold for the center clip process. The
threshold is half of the peak value over entire signal in the
buffer. The speech signals on the buffer are clipped, so that
we would reduce computation complexity. The clipped
signals are processed in autocorrelation computation. The
autocorrelation process would result two kind patterns.
There are peak-valley-peak pattern and peak-valley pattern.
The pitch value is detected based on the distance between
peaks of the peak-valley-peak pattern. The peak-valley
pattern indicates that the signals are unvoiced.
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
s(n)
33
Buffer
30 ms
sb
Signal
detector
voiced and
unvoiced
decision
svad
svad
suv
sv
cpk
Tp Characteristic
signal
Formant
codebook
FFT
sFFT
cFn
sss
Peak Code
searching
cpk
vpk
Tp
Pitch
detector
Spectral
Smoothing
Peak
codebook
sp
Decimation
Period
codebook
M
u
l
t
i
p
l
e
x
i
n
g
D
i
Segmental
quantization
Tp
vpd
cpd
Period code
searching
cpd
c(n)
Fn
Formant code
searching
Figure 1. Diagram of the Encoder
The voiced and the unvoiced signals are classified by
using the pitch detection process results. The autocorrelation
results pattern are used as reference to identify the signals
whether voiced or unvoiced. If the pattern is peak-valleypeak and if the distance between peaks is longer than 2.5 ms
but less than 20 ms, it means that the signals is voiced. Then,
if the pattern is peak-valley or peak-valley-peak with
distance between peaks is longer than 20 ms, it means that
the signal is unvoiced. The voiced and unvoiced signals
would process in the different methods. The unvoiced
signals would be process without referring the pitch period,
then the voiced signals would process based on the pitch
period.
The voiced signals are fetched on the one pitch period
that representing entire voiced signals on the buffer. The one
pitch period signals are called as characteristic signal in
waveform interpolative signal terminology [8][9]. The
length of the characteristics signals is referred as the pitch
period. The characteristic signal is quantized on its peaks
and periods by using segmental sinusoidal model. For the
unvoiced signal, the decimation process is implemented to
obtain the smaller size of signal. Then the peak and period
quantization is applied. Based on the segmental sinusoidal
model, peaks and periods information is extracted. The
processed signal would be generated by using the peaks and
periods quantization.
The peak information size is reduced by applying 10
look-up tables. The look-up table is also called as codebook.
The codebook is trained by using the peak information codevector. Large amount of the peak information code-vectors
are trained with k-means algorithm to obtain the peaks
codebook. The index number of the peak codebook is varied
from 6 to 10 to obtain the optimum process. The period
information size is also reduced by applying look-up table.
The index number of the peak codebook is also varied from
6 to 10 to obtain the optimum process. The period accuracy
has to maintain to obtain the good receiver perception on the
decoder side. Formant information is important for postfiltering process. The post-filter is arranged as four adaptive
band-pass filters to increase the formants and decrease the
valley between formants to enhance perception. The
formants location is detected by applying the FFT and
smoothing filter. Then the low-passed spectra are detected
on its peaks location. The peaks location is send to receiver
to set the center frequency on the appropriate band-pass
filter.
Parameters that have to be sent to the receiver are peaks,
periods, pitch, formants, segment, and decimation. The
peaks information needs 16-56 bits, the period information
needs 6-54 bits, pitch needs 7 bits, segment information
needs 6 bits and the formants information needs 0-14 bits.
The maximum total coded signal bits resulted for one frame
(30 ms) is 120 bits. Thus, the coded speech signals data rate
is 4 kbps.
Signal
detector
silence
Se
Ss
c(n)
D
e
m
u
l
t
i
p
l
e
x
i
n
g
cpd
cpk
Sr
Unvoiced
reconstruction
sˆ( n )
+
SF
Sr
postfilter
Voiced
reconstruction
Tp , i, D
Fn
Figure 2. Diagram of the Decoder
In decoder, the coded signal is reconstructed to obtain the
speech signal approximation with near toll quality at rate 4
kbps or bellow. If the coded signal is detected as silence, the
decoder will generate silence over a frame length i.e. 30 ms.
If, it is an existing speech signal, the decoder will generate
the speech signal depends on the type of signal. Unvoiced
signal is generated using the sinusoidal parameters. For
voiced signal, there is a special processing, that is postfiltering process. The post-filter will enhance performance of
the voiced signal by increasing the formant peaks and
reducing the valley between formant. The encoder sends
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
34
Proceedings of CITEE, August 4, 2009
speech signal information as a multiplexed parameter. In the
decoder, parameters are de-multiplexed to obtain the useful
information for signal reconstruction. Parameters would be
generated are peaks, periods information, formants
information, pitch, number of segment, and decimation.
The signal detector would identify the kind of signal. If
there is a silence, detector would not work because. The
unvoiced signal will be reconstructed if it is detected as
unvoiced. The peaks and periods are used to generate the
unvoiced signal. The next process is interpolation process
with ratio D, inverting of the decimation process on the
encoder. The voiced signal is reconstructed by generating the
characteristic signal along the 30 ms segment. The number
of characteristic signal along this segment varies between 1.5
until 12, depends on the pitch period. The characteristic
signal is generated by using the segmental sinusoidal model
from peaks and periods information. The reconstructed
signal, especially the voiced signal is passed into post-filter.
The post-filter proposed is a train of four adaptive band-pass
filters. The center frequency of each filter is changed every
30 ms segment. The center frequencies are detected by
encoder, and then they are sent to decoder. Then the comb
filter and compensation filter are applied to improve the
hearing perception quality [11][12].
The resulted reconstructed signal by sinusoidal approach
seems smoother than the original that had arbitrary form
between one peak to the consecutive peak. Nevertheless,
roughness of original signal means containing high
frequency component. Therefore, the spectral power is
reduced on the high frequency component. Unfortunately,
increasing of the number of peak would decrease the
dynamic range of period changing variation for each
segment. Thus, compression ratio would be increased to
compensate decreasing of compression ratio caused by the
number of peaks.
IV.
EXPERIMENTAL RESULTS
The proposed speech coder has been simulated in a
personal computer using a 4kbps coder program that was
specially developed by using C++ programming language.
The resulted mean opinion score (MOS) test measured for
15 Indonesian phrase in computer simulation is 3.8 (out of
5). It is tested on 46 peoples with variation on gender,
background and age. The coder complexity is less than 15
MIPS, comparable with others kind of speech coder that
needs 0.01 MIPS until 90 MIPS. It need less than 30 kB for
encoder and 10 kB for decoder.
Then, the coder is implemented on digital signal
processor starter kit TMS320VC5416, using a 4 kbps coder
program specially developed for the DSP system, based on
Code Composer Studio ver.2. Based on the experimental
results, hearing perception of the reconstructed signal is
fairly good. By using digital signal processor starter kit, the
MOS test score is 3.3 (out of 5), because of reducing the
number of codebooks. The resulted mean opinion score
(MOS) test measured for 15 Indonesian phrase in computer
ISSN: 2085-6350
simulation is 3.8 (out of 5). It is tested on 46 peoples with
variation on gender, background and age. The coder
complexity is less than 15 MIPS. It need less than 16 kB for
encoder and 3 kB for decoder. Therefore, it is expected that
the proposed low bit rate speech coder with fairly good
MOS test score and low complexity will be suitable for
voice communication and telemedicine applications during
and after disaster cases.
Figure 3. MOS test results
V.
CONCLUSION
Speech signal could be coded into 4 kbps rate and
decoded with high quality of the human perception based on
segmental sinusoidal model. The maximum MOS score is
3.8 on Indonesian words. The coder complexity for the
digital signal processor implementation is low, it needs less
than 10 MIPS.
REFERENCES
[1]
[2]
Ridkosil F., Extreme Waveform Coding, US Pattent
TF Quatiery, RJ.McAulay, Speech Transforma tions Based on a
Sinusoidal Representation, IEEE TASSP, vol. ASSP-34, no. 6, 1986
[3] RJ.McAulay, TF Quatiery, Speech, Analysis/ Synthesis Based on a
SinusoidalRepresentation, IEEE TASSP, vol. ASSP-34, no. 4, 1986
[4] T. Abe, dan M. Honda, Sinusoidal Model Based On Instantaneous
Frequency Attractor, TSALP vol 14 No.4, 2006
[5] R. Boyer, dan Abed-Meraim, K, AudioModeling Based on Delayed
Sinusoids, TSAP vol 12, No.2, 2004.
[6] CO. Etemoglu, dan V. Cuperman, V, Matching Pursuit Sinusoidal
Speech Coding, TSAP vol 11, No.5, 2003.
[7] GH. Hotho, dan RJ. Sluijter, A Narrowband Low Bit Rate Sinusoidal
Audio and Speech Coder, Company Research - Philips Electronics
Nederland, 2003.
[8] Gottesman, O. dan Gersho, A. (2000) : High Quality Enhanced
Waveform Interpolative Coding at 2.8 kbps, Proceedings of
International Conference on Acoustics, Speech, and Signal
Processing, 0-0.
[9] Gottesman, O. dan Gersho, A. (2001) : Enhanced Waveform
Interpolative Coding at Low Bit-Rate, IEEE Transactions on Speech
and Audio Processing, 9, 1-13.
[10] JR. Deller, JG. Proakis, JHL. Hansen, Discrete-Time Processing of
Speech, Macmillan Publishing Company, New York, 1993.
[11] FB. Setiawan, S. Soegijoko, Sugihartono, S. Tjondronegoro,
Perbaikan Mutu Sinyal Suara Tercampur Derau Berdasarkan Bentuk
Formant, SITIA Proceedings, Surabaya, 2006
[12] FB. Setiawan, S. Soegijoko, Sugihartono, dan S. Tjondronegoro, A
Low Bit Rate Speech Coder using Segmental Sinusoidal Model for
Disaster and Emergency Telemedicine, Journal of eHealth
Technology and Application, September 2008.
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
35
AUTOMATIC ABNORMAL WAVES DETECTION
FROM THE ELECTROENCEPHALOGRAMS OF PETIT MAL EPILEPSY CASES TO SORT
OUT THE SPIKES FP1 - FP2, THE SHARPS, THE POLYPHASE BASED ON THEIR
STATISTICAL ZEROCROSSING
Siswandari N, Adhi Susanto,
Electrical engineering UNDARIS,Electrical engineering UGM
[email protected], [email protected]
Zainal Muttaqin
neurosurgent epilepsy medical Undip
[email protected]
ABSTRACT
To explore the merits of the widespread digital signal
processing on EEG waves the standard analyses of
autocorelation, Fourier transform and zero-crossings were
applied on the seizure related waves of a large number of
epilepsy patients.
Patient consist of 267M,140F, ages 3-55 years. Clinical
status: epileptic (97M,46F: EEG with spike;39M,
28Fwithout); clinical status: nonepileptic: (93M, 41F EEG
with spike ; 38M, 25F without). Numerical data were
acquired with EEG dump diagnostic Biologic system at
Sarjito hospital Yogyakarta etc, 2002-2009. The wave
components were sorted out according to their amplitudes
and time span between zerocrossings.
The experimental results show a promising application of
the scheme, since practically all are in line with those
assessed by specialists in this field. Utilizing the hardware
and software facilities at hand, marking the st arts and ends
of abnormal waves could be done with +39µV threshold.
The zerocrossings detection could automatically
distinguished according to the 20ms–70ms time period for
the “spikes” (109M,53F) ,70ms-120ms for “sharps” (101M,
51F), and the existence of multiple peaks for
“polyphase”(57M,36F).
The research carried out so far was to find the prospect of
this digital signal processing on EEG waves to support the
doctors' work in this field.
Keywords
: EEG, spikes, sharp
I. INTRODUCTION
The clinical electroencephalographer correlates central
nervous system functions as well as dysfunction and
diseases with certain paterns of the electroencephalogram
(EEG) on an empirical basis, Obviously, this method has
been found Valuable in clinical practice
The
International
Federation
of
societies
for
electroencephalograpy and clinical Neurophysiology (
committee chaired by G.E Chatrian, see reference for
IFSECN, 1974 proposed the following definition
1. Alpha rhythm, EEG wave pattern that has
frequency 8 -13 Hz, with amplitude 20 – 200 µV.
This wave pattern can be found in almost all
normal people who are in awaked condition but
quiet.
2. Beta rhythm, wave pattern that has frequency 13
Hz, sometime in certain condition can reach 50 Hz.
Rhytmical beta activity is encountered chiefly over
the frontal and central regions, it usually does not
exceed 35 Hz.
3. Theta rhythm, wave pattern that has frequency 4 7 Hz. This wave appears in children but sometime
appears in adult who are in frustrated condition.
4. Delta rhythm, wave pattern that has frequency
under 4 Hz, sometime only appears once in two or
three seconds. This wave happens in people
sleeping tight.
The EEG utility available in a hospital have not been able to
detect spike, sharp, and polyphase automatically only in
subjectively perception and manually counting. Therefore, it
needs a program to differentiate spike, sharp and polyphase
automatically.
However, it should be noticed that a small part of EEG inter
epilepsy attack does not show any difference by using the
autocorrelation and cross correlation EEG signal. Moreover,
FFT is able to show the previous signal exploration.
II.
MODEL,
ANALYSIS,
DESIGN,
AND
IMPLEMENTATION
A. Research Material
Numerical data were acquired with EEG dump
diagnostic Biologic system at Dr. Sardjito hospital
Yogyakarta etc, 2002- by changing the file of EEG
dump diagnostic into the form of .txt. format.
B. Tools
Computer with Matlab 6.5 software using Windows XP
System
C. How To Do
The research is done with the following stages:
a. First, the collection of data cm is done by
observing the EEG qualitative diagnostic and
interpretation of its abnormal wave.
Data can be classified into:
1. Clinical status : epileptic with spike
2. Clinical status epileptic without spike
3. Clinical status : non epileptic with spike
4. Clinical status non epileptic (normal)
without spike
b. Digital data taken from the EEG biologic system of
Dr. Sardjito hospital Yogyakarta, based on Mark
Dump Utility.
To determine the wave form of spike, sharp and
polyphase in EEG, it uses zero crossing time with t0,
t1, t2, …. Etc.
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
36
Proceedings of CITEE, August 4, 2009
T1 = t1 - t 0
PSD or Power Septral Density, defined as power spectrum
quotation frequency.
T 2 = t 2 - t1
Pxx (ω ) =
T3 = t3 - t2
etc…………..(.1)
ten take the balance.
S xx ( f )
S xx (ω )
and Pxx ( f ) =
fs ...(4)
2π
III. RESULT
2.1 Interictal Paroxysmal Pattern
Spike
According to IFSECN (1974), a spike is a transient, clearly
distinguished from the background activity, with pointed
peak at conventional paper speed and a duration from 20 to
under 70 msec, the main component is generally negative,
Amplitude is variable.
The distinction from the background activity is based on
wave morphology and amplitude. In many cases, spikes
stand out against the background because of their high
voltage. If the voltage of spikes and background activity is
approximately equal, the faster character ( short duration) of
the spike is its distinctive feature.
Sharp
According to IFSECN (1974), a sharp is a transient, clearly
distinguished from the background activity, with pointed
peak at conventional paper speed and duration of 70 - 200
msec. the main component is generally negative relative to
other areas.
Jasper(1941) pointed out that the rising phase of the sharp
waveis of the same order of magnitude as in spikes, but the
descending phase is prolonged.
Figure : Spectrogram signal FP1 – FP2 for 379s.
2.2 Signal appearance by using spectrogram:
The example of signal EEG from a patient name A who is
diagnosed epilepsy petit mal taken in electrode FP1– FP2.
The purpose of spectral estimation is to explain about power
distribution which is contained in signal at certain range
frequency, according to the amount of limited data
quotation. This spectrum estimation can use to analyze
EEG’s power spectrum signal to see abnormal symptom
that happen in activity of nerve human being brain.
To see signal power spectrum clearly can be done by
dividing signal become some parts and then in each part of
the signal search their PSD. For example FP1.-FP2 signal
recorded for 6 minutes, 1.009 seconds with the quotation
frequency 256, will be separated into some parts where each
part consists of 256 data quotations, with overlapping 200
data, by using 2D image (spectrogram) the result will be
seen as figure 1.
1.
Power spectrum from a stationer random process xn,
mathematically correlating with Fourier transformation
from its row correlation, in form of normalization
frequency, this correlation can be written as:
S xx (ω ) =
∞
∑R
m = −∞
xx
( m ) e − jωm
.…..(2)
Sxx is power spectrum, Rxx is auto correlation, equivalent
above can be expressed as frequency function f (in Hertz)
by using correlation ω = 2πf/fs, with fs is quotation
frequency.
S xx ( f ) =
∞
∑R
m = −∞
ISSN: 2085-6350
xx
(m) e − 2πj f m / fs
…(3.)
The biggest figure box is power spectrum representation in
the form of 2D with abscissa asserts quotation time and
coordinate asserts frequency content with color that
represent power per set of frequency in certain time. In that
figure can be seen that frequency content with high power
happens in frequency between 0 to 30 Hz (red). When the
symptom of epilepsy appears in that frequency area become
higher (flatten red). In that figure also visible a signal
trouble that appears from power supply with frequency 50
Hz and 1.0000 Hz (orange ribbon) that appears in whole
signal.
This second is special characteristic that seen by medic in
electrode FP1 – FP2 appears 4 1.00 with very high
amplitude, 3 1.00 with average amplitude, with pattern 1.00
sharp then multi 1.00 sharp 4.00 then multi 1.00 with very
high amplitude. It approves that patient epilepsy petit mal
seen from medical science that in 1.00 seconds appear 31.00
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
37
is not be proven (or possible to be said not always). 1.00
µ
V happens in a last
with high amplitude about 65.20
second in elapsed time 00 : 03:48 – 00 :03 :49. In this
second is special characteristic of abnormal wave of petit
mal, doctor usually sees symptom clinic at first, EEG just
help to build diagnose.
Figure 4 EEG , FP1 – FP2 time elapsed 00:04:51 – 00:04:52
Tabel 1 : Zerocrossing method EEG FP1 – FP2 time elapsed 00:04:51 –
00:04:52
low
Figure 2: EEG Bipolar 19 channel
This pattern has been described as “petit mal absence
epilepsy “. Classical 3/sec spike wave complexes are widely
known
even
outside
the
community
of
electroencephalographers at FP1 – FP2 figure 1. It is
officially termed “ spike and slow wave complex “ , this
term comprises all types of spike wave complexes, which
listed separately because of markedly differing associated
clinical epileptological conditions.
Figure 3 :
00:05:00
signal FP1 – FP2
time elapsed 00:04:51 –
high
Duration
(mS)
Amplitu
de( µ V)
Code
1.00
6.00
19.53
6.80
0
1.00
21.00
34.00
50.78
7.60
4.00
poliphase
1.00
44.00
48.00
15.63
21.60
0
1.00
69.00
83.00
54.69
9.60
4.00
poliphase
1.00
93.00
114.00
82.03
15.60
4.00
poliphase
1.00
114.00
114.00
0
1.00
122.00
128.00
23.44
43.00
1.00 spike
1.00
135.00
137.00
7.81
2.40
0.00
1.00
145.00
193.00
187.50
45.00
4.00
poliphase
1.00
202.00
206.00
15.63
16.40
0
1.00
222.00
257.00
136.72
91.00
2.00 sharp
1.00
0
0
Applying Zerocrossing of Equation (1) to the detection of
epileptic spikes, sharp and poliphase . Initially, with Code
1.00 = spike, code 2.00 = sharp, code 4.00 = poliphase with
frequencies sampling fs = 256. The x axis of this output
represented the sample data n , i.e. time. The y axis of this
output represented amplitude. For table 1, low = 122.00,
high = 128.00, duration = 23,44 mS, amplitude = 43 µ V,
Code = 1.00 spike,
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
38
Proceedings of CITEE, August 4, 2009
4. CONCLUSION
1.
2.
3.
4.
The EEG wave figure shown by the recent EEG
utility can be seen from other aspect such as
autocorrelation and power spectrum dense.
The figure of spike, sharp and polyphase can be
seen in detail using zero crossing in software
automatically.
Therefore,
many
pattern
introduction might appear into sharp, spike and
polyphase.
The study can lead to the prediction of spike and
sharp waves emerge since they cannot be seen
using the ordinary EEG record.
The epilepsy sensor implant can be made in order
to plan software and hardware installation for early
warning system of an attack. The attack early
warning system can detect an attack 10 minutes
before it happens. The alarm will automatically
ring. It can happen if there is cooperation between
medical and engineering department.
REFERENCES
1 . Ernst Neidermeyer, M.D and Fernando Lopes Da Silva,
M.D, February 1982, “ Electroencephalography Basic
Principles, Clinical Applications, and Related Fields”,
Williams and Wilkins, Baltimore USA.
2 A Barreto et al.,October.1993, “Intraoperative Fokus
localization System based Spatio-Temporal EcoG
[]
[]
[3]
Analysis”, Proc. XV Amnual Intl.Conf of the IEEE
Engineering in Medicine and Biology Society.
Mahar Mardjono, 1992, Simposium Diagnosis dini
Epilepsi, MDK/ Vol. 11. No. 01/Januari 1992, Jakarta.
. Alan V. Oppenheim and Ronald W. Schafer, 1994,
“Digital Signal Processing”, Prentice Hall of India,
New Delhi.
. Erik Bolviken, Geir Storvik, Gutorm Hogasen and Pal
G. Larsson, 2004, “Spike detection in EEG through
random line segments”, The project has been
performed at the Norwegian Computing Center
supported by Royal Norwegian Council for Scientific
and Industrial Research.
Lim Shih Hui , 2003, “Interictal and Ictal Epileptiform
Discharges”, National Neuroscience Institute,
Singapore.
Hamid Hassanpour, Moestefa Mesbah and Boualem
Boashash, 2004,” EEG Spike detection Using TimeFrequency Signal Analysis”, 2004 IEEE.
Shoichiro Nakamura, 1993, “Applied Numerical
methods ”, Prentice-Hall International, Inc USA
10.John G. Webster , 1995, “Medical Instrumentation
Application and Design” John Wiley & Sons,Inc, New
York.
[4]
[5]
[6]
[7]
[8]
.
ISSN: 2085-6350
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
39
Studies on the Limitation of
Fighter Aircraft Maneuver
Caused by Automatic Control Design
Okky Freeza Prana Ghita Daulay, Arwin Datumaya Wahyudi Sumari
Departemen Elektronika, Akademi Angkatan Udara
Jl. Laksda Adisutjipto, Yogyakarta – 55002
[email protected], [email protected]
Abstract— In 1950, fighter aircraft were not able to make a
shape maneuver in very high speed. Later, it became possible
in 1980 caused by progress in technology. However, the
maneuver was unstable. The control systems couldn’t perform
perfectly and became very expensive when it applied to fighter
aircraft. In this paper, we analyzed a non-linear and a linear
model of fighter aircraft. We used Sweden JAS-39 fighter
aircraft. As we know that this fighter aircraft can perform
shape maneuver in supersonic speed during flight. To perform
this kind maneuver, it needs good control system to assist pilot
when “drive” the aircraft. Based on this problem, we design a
linear controller based on a non-linear model of fighter
aircraft and applied it to the dynamic non-linear model of
fighter aircraft. Finally, we inspected it performance based on
the angle of attack and the sideslip angle of fighter aircraft.
sensor information, the control system computes the control
surface deflections to be produced. This information is sent
through electrical wires to the actuators located at the
control surfaces, which in turn realize the desired
deflections. This is known as fly-by-wire technology. Based
on this technology, we derive a non-linear and a linear
approximation model from the dynamic model of fighter
aircraft. We used Sweden Air Force fighter aircraft JAS-39
Gripen as fighter aircraft model that shows in Fig. 1 for this
study. Here p s , q s , rs are stability-axes angular velocity, α
is angle of attack, and β is sideslip angle of fighter aircraft.
Some parameter on this model we assumed to be constant.
Keywords—automatic control system, fighter aircraft, nonlinear and linear models
I.
INTRODUCTION In 1950, fighter aircraft were not able to make a shape
maneuver in very high speed. Later, it became possible in
1980 caused by progress in technology. However, the
maneuver was unstable. The control systems couldn’t
perform perfectly and became very expensive when it
applied to fighter aircraft. In this study, we analyzed a nonlinear and a linear model of fighter aircraft. We design a
linear controller based on a non-linear model of fighter
aircraft and applied it to the dynamic non-linear model of
fighter aircraft. Finally, we inspected it performance based
on the angle of attack and the sideslip angle of fighter
aircraft. When inspected the aircraft maneuver ability, we
also inspected the possibility using linear controller to nonlinear model. We assumed, with linear controller that built
from linear model, it can control the non-linear model of the
aircraft as good as non-linear controller. Also, by using
linear controller, we can save more time and cost to build
and produce it at near future.
Figure 1. Fighter Aircraft Dynamic Model.
A. Non­Linear Model In a non-linear model, angle of attack and sideslip angle of
fighter aircraft assumed to be the variable of model. As
system input, we put stability-axes angular velocity. We
collect the fighter aircraft model parameter from the
previous section and write the result in a form suitable for
control systems. Equation (1) is the fighter aircraft force
equations.
II.
FIGHTER AIRCRAFT PRIMER
During the early years of flight technology, the pilot
was in direct control of the aircraft control surfaces. These
where mechanically connected to the pilot's manual inputs.
In modern aircraft, the pilot inputs are instead fed to a
control system. Based on the pilot inputs and available
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
40
Proceedings of CITEE, August 4, 2009
・ 1
V = ( − D + Fcosαcosβ + mg1 )
m
・
1
α = q − (pcosα + rsinα )tanβ +
( − L − Fsinα + mg 2 )
mVcosβ
・
1
β = psinα − rcosα +
(Y − Fcosαsinβ + mg 3 )
mV
.
(1)
g1 = g( − cos α cos β sin θ + sin β cos θ sin φ + sin α cos β cos θ cos φ )
g 2 = g(cosαcosθcosφ + sinαsinθ )
g 3 = g(cosβcosθsinφ + sinβcosαsinθ - sinαsinβcosθcosφ )
Equation (2) is the fighter aircraft dynamic non-linear
model of fighter aircraft that we will use for designing the
controller.
・
β = − rs +
1
mV
・
rs = u 3
1
( − L( α ) - Fsin α + mg 2 )
mVcos β
(4)
where A and b are given by
F − mgsin θ
mV
A 12 =
)
A 21 = 0
mgcos θ cos φ − L( α )
mV
A 22 = − (
b11 = 1
b12 = 0
b 21 = 0
b 22 = −1
F − mgsin θ
mV
− ps
)
where the parameters value which we used in this research
that named body mass, force, gravity acceleration, and
velocity of fighter aircraft are
.
m = 9,100 [kg]
2
g = 9.8 [m/ s ]
III.
(Y( β ) - Fcos α sin β + mg 3 )
(2)
where
g 2 = g(cos α cos θ cos φ + sin α sin θ )
g 3 = g(cos β cos θ sin φ + sin β cos α sin θ - sin α sin β cos θ cos φ )
Later, we created the fighter aircraft non-linear control
system inputs based on bakstepping method and bode
diagram based on (2). Backstepping method is a method
that built from subsystems that radiate out from an
irreducible subsystem that can be stabilized using some
other method. Because of this recursive structure, the
designer can start the design process at the known-stable
system and feedback new controllers that progressively
stabilize each outer subsystem. The process terminates
when the final external control is reached.
A. Linear Model In linear model, we assumed angle of attack and sideslip
angle is zero and roll angular velocity assumed to be
constant during calculation. This made the calculation of
problem easier to solve. Equation (3) is basic model of
linear equation.
・
x = Ax + bu
⎡・⎤ ⎡A A ⎤⎡α⎤ ⎡b b ⎤⎡q ⎤
⎢α・⎥ = ⎢ 11 12 ⎥⎢ ⎥ + ⎢ 11 12 ⎥⎢ s ⎥
A
A
b
b
r
β
⎣⎢β⎦⎥ ⎣ 21 22⎦⎣ ⎦ ⎣ 21 22⎦⎣ s ⎦
A11 = − (
where the contribution due to gravity are given by
・
・
p s = u1 q s = u 2
・
α = q s − p s tan β +
Equation (4) is the state variable result from
linearization of the fighter aircraft dynamic equation
according to (3).
F = 54,000 [N]
V = 165.811 [m/s]
FIGHTER
AIRCRAFT CONTROLLER
We designed non-linear controller based on a non-linear
model and a linear controller based on linear model. We
used Scilab/Scicos program to design and simulate the
system. Scilab/Scicos program is a numerical computation
software package similar to MATLAB/SIMULINK that has
been released by Institut National de Recherché en
Informatique et Automatique (INRIA) from France since
1990.
A. Non­Linear Controller We referred to [1] when designed non-linear controller
using Backstepping method based on non-linear model.
According to the non-linear model in (1), (5), (6), and (7)
are non-linear controller inputs using Backstepping method.
u1 = k ps (ps
ref
u 2 = − k α,2 (q s + k α,1 (α − α
(5)
− ps )
ref
) + f α (α
u 3 = k β,2 ( − rs + k β,1β +
ref
gcosθsinφ
V
, y α ))
(6)
)
(7)
Fig. 2 is fighter aircraft non-linear model bode diagram
using Backstepping method.
(3)
ISSN: 2085-6350
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
41
k = −r
+
rs −
ref
kβ ,1
+
ps
ref
+
−
k ps
−
1
s
u3
1
s
・
-1
β
+
u2
−
α
・
Equation (8) is fighter aircraft linear model equation
using enlarged system method that we used to solve the
matrix problem during build the feedback gain.
・
⎡ v(t) ⎤
x = A x (t) + bu(t) +
⎢⎣ r(t) ⎥⎦
(8)
where
b=
⎡b ⎤
⎢⎣0 ⎥⎦
⎡ F − mgsinθ mgcosθgcos − L(α ref )
ref
− ps
)
⎢− (
mV
mV
⎢
0⎤
F − mgsinθ
=⎢
)
0
−
(
0⎥⎦ ⎢
mV
⎢
−1
0
⎢
0
−1
⎣
⎡1 0 ⎤
⎢0 − 1⎥
=⎢
⎥
⎢0 0 ⎥
⎢⎣0 0 ⎥⎦
0
0
0
0
According to linear model in (8), I made linear model
input as shown in (9).
u = kx
(9)
where k is feedback gain. We used Riccati Equation as
shown in (10).
A T P + PA − r − 1 PBB
T
P+ Q = 0
(10)
This equation used to build feedback gain which means the
controller in linear model and applied it to the linear model
inputs as shown in (11).
u = −r
−1 T
B Px
Figure 3. Linear System
A. MANEUVER AND SIMULATION In this session, we combine three control models and
controllers to inspect each combination performance.
They are:
• Non-linear model with non-linear controller.
1
s
B. Linear Controller We used linear quadratic regulator method when
designed linear controller based on linear model of fighter
aircraft. Meanwhile, to design a servo system that suitable to
appointed angle of attack and sideslip angle, we used the
Enlarged system method.
⎡A
⎢⎣− c
Equation (12) is feedback gain from linear model. Fig. 3 is
fighter aircraft linear model bode diagram based on (8).
Here, roll angular velocity assumed to zero because it has
less effect both in angle of attack and sideslip angle
maneuver.
α
Figure 2. Non-linear System
A=
(12)
tan β
+
qs ref −
β
−1 T
B P
•
Linear model with linear controller.
•
Non-linear model with linear controller.
These three-combination-and-simulation goals was to
inspect the differences on each combination, search the best
combination for the model, and look forward about
possibility to use new combination between model and
controller. Meanwhile, we also figure out each combination
effect to fighter aircraft maneuver during flight.
Fig. 4 and Fig. 5 (a) is non-linear model with non-linear
controller, (b) is linear model with linear controller, and (c)
is non-linear model with linear controller simulation result.
Fig. 6 is non-linear model with linear controller which has
been increased in feedback gain. In this result from top to
below showed roll angular velocity, angle of attack targeted
value, sideslip angle targeted value, angle of attack result,
⎤
0⎥ and sideslip angle result.
⎥
Fig. 4 described the simulation from fighter aircraft
0⎥
⎥ maneuver without roll maneuver. Here we inspect each
0⎥
controller design effect to fighter aircraft maneuver in
0⎥⎦
simply maneuver environment. Fig. 4 (a) and (b) mostly
showed similar movement simulation result especially in
angle of attack (AoA) or pitch angle. But, Fig. 4 (b) showed
that sideslip angle grew bigger. It may be said that the
adjustment of the control loop for the sideslip angle is more
necessary. In (c) the result was able to made similar
movement especially in angle of attack (AoA) or pitch
angle. But, it has too much noise with it. It was started when
pilot gave the input to aircraft and continued even pilot did
not give any input. Meanwhile, sideslip angle result turned
unstable at all. Even pilot did not give any input, result
showed movement that indicate aircraft did not work
normally. When we increased input continuously, angle of
attack (AoA) showed less noise on it. Meanwhile, noise on
sideslip angle and last combination (non-linear model with
linear controller) getting bigger continuously.
(11)
refer to (11), the feedback gain for this model is
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
42
Proceedings of CITEE, August 4, 2009
ps
ref
α ref
+
1
s
−
β ref
+
−
1
s
ps
Gain
(
k)
α
qs
Control
rs
System
β
Fig. 5 described the simulation from fighter aircraft
maneuver with roll maneuver. We inspect each controller
from first and second simulation, this combination did
not show good result. We increased feedback gain
controller by increased heaviness for the state (Q) to
stabilize system. Fig. 7 (a) gave the best result among them.
When I set Q = diag (100, 100, 1, 1), angle of attack (AoA)
and sideslip angle result showed closer to input. If we grew
feedback gain, angle of attack (AoA) result getting more
unstable and could not reach the same value as input.
Meanwhile, sideslip angle result showed better because it
could reach the same value as input. Even after we set the
heaviness for the state (Q) extremely when designed linear
controller, it couldn’t improve the result at all when applied
to non-linear model.
Therefore, it might be said that linear controller that built
from linear model, was not able to control the non-linear
model of fighter aircraft and fighter aircraft maneuver
performance received big influence by it control system.
V. CONCLUDING REMARKS AND FURTHER WORKS On this study, we guessed it whether cannot control the
non-linear model of the fighter in a linear controller.
In addition, we thought it to be useful when we
implemented it because the result did not become
expensive, less design time, and many other positive results.
However, in this study, we simulated it and inspected it, but
our assumed result did not appear.
Finally, there was only two ways controller design this
far. First, non-linear controller designed by it own system
and linear controller designed by it own system. Each
controller can only control it own system. Also, it is better
to use non-linear controller for fighter aircraft non-linear
model than use linear controller that built from linear
model.
As reference for next research is make a computer
program to control servo or stepper based on this research.
This is to prove that the simulation result can be
implemented on reality or not.
ISSN: 2085-6350
design effect to fighter aircraft maneuver in complex
maneuver environment. Fig. 5 (a) and (b) showed similar
result with first simulation result. In angle of attack (AoA)
or pitch angle result, Fig. 5 (a) and (b) showed similar result
for input and output. It described that was no problem
during control the aircraft. Aircraft could perform maneuver
as well as input that was given by pilot. Meanwhile, noise
that occurred in sideslip angle was getting bigger that result
in first simulation. It showed that more complex input that
pilot give, the result will become more complex. Fig. 5 gave
same result as first simulation.
Fig. 6 described the simulation from combination of
non-linear model and linear controller. According to result
For our further works, we would like to try another method
to resolve all the problems that appear during research. As
an example, we would like to rebuild non-linear controller
using Lyapunov method. Lyapunov method allow us to
analyze the stability from a system, in this case is fighter
aircraft stability. By using lyapunov method, we can
compare the result between it when applied to build
controller and the controller using Backstepping method.
From this result we can compare which one has the best
result, linear controller, non-linear controller using
Backstepping method, or non-linear controller using
Lyapunov method. We also will try new model with same
research. Next model for research is F-16 Fighting Falcon.
We will simulate same simulation and compare the result
with JAS-39 Gripen result.
ACKNOWLEDGMENT First of all, we want to thank our mighty God. Also we want
to thank National Defense Academy of Japan Department of
Electronic and Electrical Engineering, especially Control
System Research Group for hereby giving me the
opportunity to perform research. We also want to thank my
supervisor Professor Sasaki Keigo for guidance and
expertise within non-linear and linear control system,
respectively. We also want to thank Governor of Indonesian
Air Force Academy and all member of Electrical
Department for support during this research.
[1]
[2]
[3]
[4]
[5]
REFERENCES O. Harkegard, “Flight Control Design Using Backstepping”,
Linkopings University, 2001
M. Najafi & R. Nikoukhah, "Modeling and simulation of differential
equations in Scicos", pp.177-185, The Modelica Associations,
September 4 th − 5 th , 2006
B. Etkin & L.D. Reid, "Dynamics of flight: stability and control",
AAIA, 1996.
Y.C. Yamaguchi, "Fundamentals of Aerospace Engineering", pp.5158, 2003-2008
O.F.P.G.
Daulay,
"制御器設計に起因する戦闘機の旋回性能限界に関する研究”,
National Defense Academy of Japan, 2009
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
ps
43
ref
[ rad / s ]
α
ref
[rad ]
β
ref
[rad ]
α
[rad ]
β
[rad ]
(a) ΣN
(b) ΣL
(c) ΣNL
Figure 4. Simulation result without roll maneuver
ps
ref
[ rad / s ]
α
ref
[rad ]
β
ref
α
[rad ]
[rad ]
β
[rad ]
(a) ΣN
(b) ΣL
(c) ΣNL
Figure 5. Simulation result with roll maneuver
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
44
Proceedings of CITEE, August 4, 2009
ps
ref
[ rad / s ]
α
ref
[rad ]
β
ref
[rad ]
α
[rad ]
β
[rad ]
(a) Q=diag(100,100,1,1)
(b) Q=diag(1k,1k,1,1)
(c) Q=diag(10k,10k,1,1)
Figure 6. Simulation result with increased feedback gain (ΣNL)
ISSN: 2085-6350
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
45
Feature Extraction and Selection on Osteoporosis
X-Ray Image for Content Based Image Retrieval
(CBIR) Purposes
Usman Balugu1,2, Ratnasari Nur Rohmah1,3
1
Jurusan Teknik Elektro, UGM
[email protected], [email protected]
2
Politeknik Bandung, 3UMS Surakarta
Nurokhim
PTLR, Batan
City PostCode, Indonesia
[email protected]
Abstract— This paper presents results on experiment in
extracting feature of osteoporosis X-Ray image and finding the
right feature or features to perform as descriptor in ContentBased Image Retrieval (CBIR) Systems. Consider that
descriptor of each image will represent the content of each
image, the decision which features of image will be the best
representation of image content, is very important. Data in this
experiments are digitized X-Ray proximal femur images
already known osteoporosis level based on Index Singh (grade
3 – grade 6) due to simple verification on classification. We
used statistical approach and log-polar wavelet transform for
extracting textural feature from image. On statistical
approach, we use grey-level co-occurance matrix (GLCM)
methods to extract feature and used Contrast and Uniformity
as quantity measured. We use Norm-1 as energy wavelet
signature measured. Experiment on feature selection conduct
by applying feature vector on classification process.
Experiment’s results shows that two dimension feature vector
using combination of Contrast on GLCM and Norm-1 on
approximation component of log-polar wavelet transformed
image, is the best choice image descriptor’s. Log-polar
transforms proven effective in reducing wavelet energy error
of rotated and scaled image. Improvement in classification
accuracy also shown that, the use of log-polar transform prior
to wavelet transform is useful in classification study.
Keywords—osteoporosis, descriptor, GLCM, log-polar,
wavelet
I.
INTRODUCTION
Application computer in telemedicine has improved
medical service in healthcare area. One of the applications
is in applying databases on medical information such as
medical record or medical image. Recently, compare to
common method textual-base image retrieval, CBIR is
considered as the more appropriate choice method for image
retrieval in medical image databases. This method retrieves
images from database using information directly derived
from the contents of image themselves, rather than from
accompanying text or annotation [1].
As a basic principle of CBIR, images are internally
represented by numerical features, which are extracted
directly from image. These features are stored in the
database, as a descriptor, along with the images, for rapid
access. At retrieval time, the user presents a sample image,
and the system computes the numerical features, compares
them to those stored in the database, and returns all images
with similar features. It is obvious that the quality of
response depends on descriptor and the distance or
similarity measure that is used to compare features from
different images.
This paper presents results on experiment in extracting
feature of osteoporosis X-Ray image and choosing the right
feature or features to perform as descriptor in ContentBased Image Retrieval (CBIR) Systems. Consider that
descriptor of each image will represent the content of each
image, the decision which features of image will be the best
representation of image content, is very important. We used
statistical approach Grey level co-occurance matrix
(GLCM) and log-polar wavelet transform for extracting
image feature. Selection on image feature is conduct by
choosing feature vector gives optimum accuracy on image
classification test. Data in this experiments are digitized XRay proximal femur images already known osteoporosis
level based on Index Singh (grade 3 – grade 6) due to
simple verification on classification validity.
II.
FEATURES EXTRACTION AND SELECTION
A. Textural Feature on Osteoporosis X-Ray Image
Osteoporosis is characterized by an absolute decrease in
the amount of bone to a level below that required for
mechanical support of normal activity and by the occurrence
of non-traumatic skeletal fracture [2]. Bone structure can be
estimated by observing the change of trabecular pattern in
proximal femur radiograph. The observation of trabecular
pattern change for diagnosis of osteoporosis was first
proposed in the 1960s using radiographs of proximal femur.
The diagnosis was known as Singh Index grading system.
In this research, we used this structure pattern as image
feature. We used 4 grades of Singh indexed image (grade 3
to grade 6) from women patients between 45 – 65 years of
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
46
Proceedings of CITEE, August 4, 2009
age. Image feature extraction is done by image textural
analyzing which will produce feature vector of each image.
This feature vector is used as descriptor on the image
retrieval systems. We used statistical approach and log-polar
wavelet transform for extracting textural feature from image.
B. Feature Extraction Using GLCM
On statistical approach we use GLCM methods to extract
feature. As X-Ray image represent as grey-level image, this
method is considered as suitable method. From an image, f,
with L possible intensity levels, GLCM is a matrix G whose
element g ij is the number of times that pixel pairs with
intensities zi and z j occur in f , in position specified by an
operator Q , where 1 ≤ i , j ≤ L [3]. The operator Q is an
operator that defines the position of two image pixels
relative to each others. The total number, n, of pixel pairs
that satisfy Q is equal to the sum of the elements of G.
Then, the quantity
g ij
pij =
(1)
n
given N × N image f ( x , y ) is transforms into polar form
S × [N 2] image p(a, r) and then logarithm functions are
applied to this polar form to obtain S × R log-polar
image lp ( i , j ) .
Polar form S × [ N 2 ] image p ( a, r ) of N × N image
f ( x , y ) is compute as follows:
⎛⎡N ⎤ ⎡
⎛ 2πa ⎞⎤ ⎞⎟ (5)
⎛ 2πa ⎞⎤ ⎡ N ⎤ ⎡
p (a, r ) = f ⎜⎜ ⎢ ⎥ + ⎢r cos⎜
⎟⎥
⎟⎥, ⎢ ⎥ − ⎢r sin ⎜
⎝ S ⎠⎦ ⎟⎠
⎝ S ⎠⎦ ⎣ 2 ⎦ ⎣
⎝⎣ 2 ⎦ ⎣
for a = 0 , L , S − 1 , and r = 0, L , [N 2] − 1 , and S × R logpolar image lp(i, j ) is computed as follows:
⎛ ⎡ log 2 ( j + 2 )
lp ( i , j ) = p ⎜⎜ i , ⎢
⎝ ⎣ log 2 ( R + 2 )
⎡N
⋅ .⎢
⎣2
⎤ ⎤ ⎞⎟
⎥⎥ ⎟
⎦⎦ ⎠
(6)
for i =0,L,S −1, and j = 0, L , R − 1 .
is an estimate of the probability that a pair of points that
satisfy Q will have value zi , z j . These probabilities are in
(
)
the range [0, 1] and their sum is 1:
K K
∑ ∑ pij = 1
(2)
i =1 j =1
where K is the row (or column) dimension of square matrix
G.
Because G depends on Q, the presence of intensity
texture pattern can be detected by choosing an appropriate
position operator and analyzing the elements of G. In this
research position operator defined as “one pixel immediately
to the right”. As for descriptor that useful for characterizing
the content of G, we choose two descriptors, Contrast and
Uniformity. Contrast is a measure of intensity contrast
between a pixel and its neighbor over the entire image [3].
The range of values is 0 (when G is constant) to (K – 1)2.
Contrast = ∑ ∑ (i − j )2 pij
K K
i = i j =1
K K
(b) Image (a) rotated
by 10 degree
(c) Image (a) scaled by
110%
(d) Log-polar
transformed of (a)
(e) Log-polar
transformed of (b)
(f) Log-polar
transformed of (c)
(3)
Uniformity (also called Energy) is a measure of uniformity
in the range [0, 1]. Uniformity is 1 for a constant image.
Uniformity = ∑ ∑ pij2
(a) Original image
Figure 1. Original, rotated, and scaled image and their log-polar
transformed.
(4)
i =1 j =1
Wavelet Transform for textural feature extraction
C. Feature Extraction Using Log-Polar Wavelet
Transforms
Log Polar Transform
Log-polar transform is used to eliminate the rotation
and scale effect in the input image by converting the image
into corresponding log-polar image (Fig 1). Such log-polar
image is rotation invariant and nearly scale invariant [4].
This transform algorithm is divided into two major steps. A
ISSN: 2085-6350
Discrete Wavelet Transform (DWT) useful for image
analysis due to wavelets having finite duration which
provides both the frequency and spatial locality and
efficient implementation [4].
The hierarchical onedimensional DWT uses pair of high-pass and low-pass
filters derived from wavelet functions to decompose the
original signal into two subbands: details and
approximation, respectively. In 2D-DWT analysis, an
image is split into an approximation and three detail images.
The approximation image is then itself split into a second-
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
47
level approximation and detail images, and the process is
recursively repeated. The tree details images as well as the
approximation images can also be split.
The standard 2D-DWT can be described by a pair of
quadrature mirror filters (QMF) H and G [5]. The filter H is
a low-pass filter with a finite impulse response denoted by
h(n). And, the high-pass G with a finite impulse response is
defined by:
g ( n ) = ( − 1) n h (1 − n ), for all n.
(7)
keep one column out of two
keep one row out of two
The low-pass filter is assumed to satisfy the following
conditions for orthonormal peresentation:
for all j ≠ 0
∑ h(n)h(n + 2 j) = 0,
(8)
convolve with filter X
Figure 2. Decomposition of a discrete image
image by using the conjugate filters H and G.
n
∑ h(n )
2
(9)
=1
n
∑ h(n) g (n + 2 j ) = 0,
for all j
(10)
n
Forward 2D-DWT using Fast Wavelet Transform
(FWT) algorithm of a N × M discrete image x up to level
p + 1 is recursively defined in terms of the coefficients of
level p as follows:
(11)
C4pk+,1(i , j ) =
h(m )h(n )Ckp, ( m + 2i , n + 2 j )
∑∑
m
n
C4pk++11, ( i , j ) = ∑∑ h(m )h(n )Ckp, ( m + 2i , n + 2 j )
(12)
C4pk++12, ( i , j ) = ∑∑ h(m )h(n )Ckp, ( m + 2i , n + 2 j )
(13)
C4pk++13, ( i , j ) = ∑∑ h(m)h(n )Ckp, ( m + 2i , n + 2 j )
(14)
m
m
m
Ckp into four quarter-size
n
n
n
Log-polar Wavelet Transform
Wavelets transforms provide good multiresolution
analytical tools for texture analysis and classification [6][7].
Experimental results show that this approach can achieve a
high accuracy rate [4]. However, this approach assumes
that the texture images have the same orientation and scale,
which is not realistic for most practical application. A
number works address the problem on rotation and scale on
image. One on these works that shows promising results is
using log-polar transforms on image [4][8] and combined
wavelet packet transforms. Considered that image data in
this research are obtained from scanning X-Ray analog
image, which usually subject to certain random skew angles
and scaling, we use this method prior to wavelets transforms
on image texture-based feature extraction [7] (Fig. 3). As
comparison, we also build feature extraction systems
without applying log-polar transform (Fig. 4). Quantities
calculated as features in this approach are the norm-1 e from
all component of wavelet decomposed image
(approximation, vertical, horizontal, and diagonal
component); e A , ev , eH , eD .
where C0 ( i , j ) = x( i , j ) is given by the intensity levels of the
0
e=
image x.
1 N
∑ Ck
N 2 k =1
2
(15)
In practice, the image x has only a finite number of
pixels. Different methods such as symmetric, periodic or
zero padding could be used for boundary handling. At each
p
step, we decompose the image Ck into four quarter-size
images
C4pk+1 ,
C4pk++11 ,
C4pk++12 , and C4pk++13 . This
decomposition algorithm can be illustrated by the block
diagram in Fig. 2.
Figure 3. Feature extraction using GLMC and Log-polar Wavelet
Transform
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
48
Proceedings of CITEE, August 4, 2009
III.
EXPERIMENT AND RESULTS
A. Experiment and Result in Log-polar Transform effect in
Wavelet Energy Signature
Figure 4. Feature extraction using GLMC and Wavelet Transform
D. Feature Selection
As mention in sub section above feature extraction for
image is conduct using GLCM approach and wavelet
transforms (Fig. 3 and Fig. 4). That is means that for each
image we have six quantities measured as image feature.
Not all of the measure features above are considered as
descriptor. In this paper, feature vector as descriptor is a
two-dimension vector using combination of two features,
one from statistical approach and another from wavelet
transform. From these eight combinations we define which
feature vector as descriptor by trial and error (Fig. 5).
Experiment in log-polar transform is conduct to ensure
positive impact of this transform in feature extraction. As
mention in sub section above that image data in this research
are obtained from scanning X-Ray analog image, which
usually subject to certain random skew angles and scaling.
Experiment results shows that log-polar transform reducing
error in wavelet energy measured on rotated image (Fig. 6).
On the other hands, this transform doesn’t that much
significant effect on scaled image, especially for
approximation image (Fig. 7). However, generally this logpolar transform has positive effect in reducing MSE of
wavelet energy signature from all decomposed image.
(a) Approximation image
(b) Horizontal detail image
(c) Vertical detail image
(d) Diagonal detail image
Test on image classification is conduct on some sample
data using each combination of two features. A simple
minimum distance classifier method was chosen for this
testing process. Feature vectors giving optimum accuracy
on classification is selected as descriptor.
Figure 6. Mean Square Error of wavelet energy measured
from rotated image.
(a) Approximation image
(b) Horizontal detail
(c) Vertical detail
(d) Diagonal detail
Figure 5. Flowchart on feature vector selection process
Figure 7.
Mean Square Error of wavelet energy measured
from scaled image
ISSN: 2085-6350
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
49
B. Experiment and Result in Feature Selection and impact
of the log-polar transform.
Experiment on feature selection shows that only four
feature
vectors: [e A Contrast ]T ,
[e A Uniformity ]T ,
and [eD Uniformity]T , for system using logpolar transform prior to wavelet transform, that result in
clustering in classification process (Fig. 8). In systems not
applying log-polar transform, four feature vectors:
[e A Contrast ]T , [e A Uniformity ]T , [eH Uniformity]T , and
[eV
[eV
Uniformity ]T ,
wavelet transform is the best choice for image descriptors.
This result gives higher accuracy rate 86.25% of all classes
compare to accuracy rate 83.75% for the same feature
vectors in system without applying log-polar transform.
These facts also show that applying of log-polar transforms
has its benefits. Table I and Table II shows classification
accuracy based on feature vectors.
TABLE I.
CLASSIFICATION ACCURACY ACHIEVED ON
CLASSIFICATION PROCESS USING FOUR FEATURE VECORS DERIVED FROM
GLMC AND LOG-POLAR WAVELET TRANSFORM FEATURE EXTRACTION
Uniformity]T
shows clustering in classification process
(Fig. 9). Boundary lines are performed based on feature
vectors (descriptor) of 37 images in databases.
Classes
Grade 3
Grade 4
Grade 5
Grade 6
⎡ eA ⎤
⎢
⎥
⎣Contrast ⎦
(a) Feature vector:
(b) Feature vector:
eA
⎡
⎤
⎢
⎥
⎣Uniformity ⎦
eV
⎡
⎤
⎢
⎥
⎣Uniformity ⎦
(d) Feature vector:
(a) Feature vectors:
(c) Feature vectors:
Grade 3
Grade 4
Grade 5
Grade 6
eD
⎡
⎤
⎢
⎥
⎣Uniformity ⎦
Figure 8. Decision boundary of minimum distance classifier based on
feature vectors from feature extraction process
applying log-polar wavelet transform.
⎡ eA ⎤
⎢
⎥
⎣Contrast ⎦
(b) Feature vectors:
eA
⎡
⎤
⎢
⎥
⎣Uniformity ⎦
eH
⎡
⎤
⎢
⎥
⎣Uniformity ⎦
(d) Feature vectors:
eV
⎡
⎤
⎢
⎥
⎣Uniformity ⎦
Figure 9. Decision boundary of minimum distance classifier based on
feature vectors from feature extraction process
applying wavelet transform
eA
⎡
⎤
⎢
⎥
⎣Uniformity ⎦
eV
⎡
⎤
⎢
⎥
⎣Uniformity ⎦
eD
⎡
⎤
⎢
⎥
⎣Uniformity ⎦
91.67
70.00
83.33
100.00
75.00
80.00
83.33
100.00
66.67
60.00
66.67
100.00
75.00
30.00
25.00
100.00
Classification accuracy (%) for feature vectors:
⎡ eA ⎤
⎢
⎥
⎣Contrast ⎦
eA
⎡
⎤
⎢
⎥
⎣Uniformity ⎦
eH
⎡
⎤
⎢
⎥
⎣Uniformity ⎦
eV
⎡
⎤
⎢
⎥
⎣Uniformity ⎦
91.67
60.00
83.33
100.00
75.00
70.00
83.33
100.00
83.33
30.00
25.00
33.33
75.00
30.00
25.00
100.00
IV.
CONCLUSION
Combination on log-polar wavelet energy signature and
contrast on GLCM as two-dimension vector descriptor, give
sufficient accuracy for the CBIR purposes. Log-polar
transforms proven effective in reducing wavelet energy error
of rotated and scaled image. Improvement in classification
accuracy shown that, the use of log-polar transform prior to
wavelet transform is useful in classification study.
However, only small database used for verification in
this research, and most of data has already confirmed its
index Sigh. Further works would be applying bigger
database and more indistinct indicated data to find the best
image feature to represent of image content as descriptor.
Moreover, feature vector dimension would probably need to
be increased in larger database.
ACKNOWLEDGMENT
F. A. Author thanks to Mr. J. Tjandra Pramudityo for
giving copy on data of osteoporosis X-Ray images used in
this research.
REFERENCES
[1]
Among all those combination, features vectors consist
of Contrast on GLCM and Norm-1 on approximation
component [e A Contrast ]T in system using log-polar
⎡ eA ⎤
⎢
⎥
⎣Contrast ⎦
TABLE II.
CLASSIFICATION ACCURACY ACHIEVED ON
CLASSIFICATION PROCESS USING FOUR FEATURE VECORS DERIVED FROM
GLMC AND WAVELET TRANSFORM FEATURE EXTRACTION
Classes
(c) Feature vector:
Classification accuracy (%) for feature vectors:
El-Naqa, I., Yang, Y., Galatsanos, N. P., & Wernick, M. N., “A
similarity learning approach to content-based image retrieval:
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
50
[2]
[3]
[4]
[5]
[6]
Proceedings of CITEE, August 4, 2009
Application to digital mammography”, IEEE Transaction on
Medical Imaging, vol. 23, no. 10, pp. 1233-1244, 2004.
H.W. Wahner and Fogelman I., The Evaluation of
Osteoporosis: Dual Energy X-ray Absorptiometry in Clinical
Practice, Martin Dunitz Ltd., London, 1994.
R. C. Gonzalez and R. E. Woods, Digital Image Processing,
3rd ed., Pearson Education, New Jersey, 2008.
Chi-Man Pun and Moon-Chuen Lee, “Log-polar Wavelet
Energy Signatures for Rotation and Scale Invariant Texture
Classification”, IEEE Transactions on Pattern Analysis and
Machine Intelligence, vol. 25, no. 5, pp. 590-603, May 2003.
Stéphane Mallat, A Wavelet Tour of Signal Processing, 2nd
ed., Academic Press, 1999.
ISSN: 2085-6350
[7]
[8]
[9]
T. Chang, C.C. J. Kuo, “Texture Analysis and Classification
with Tree-Structured Wavelet Transform”, IEEE Transaction
on Image Processing, vol.4, pp. 1549-1560, Nov. 1995.
A. Laine and J. Fan, “Texture Classification by Wavelet
Packet Signatures,” IEEE Transaction on Pattern Analysis
and Machine Intelligent, vol. 8, pp. 472 – 481, July 1986.
Ratnasari N. R, Lukito E. N, Thomas Sri W., Adhi Susanto,
Nurokhim. “Design and Preliminary Result on Content-Based
Image Retrieval (CBIR) System for Osteoporosis X-Ray
Image Database.”, Proceeding of International Conference on
Rural Information and Communication Technology 2009, pp.
199–202, July 2009
Conference on Information Technology and Electrical Engineering (CITEE)
51 Proceedings of CITEE, August 4, 2009
The Implementation of Turbo Encoder and Decoder
Based on FPGA
Sri Suning Kusumawardani * and Bambang Sutopo †
* Dept. of Electrical Engineering, Fac. of Engineering, Gadjah Mada University, Jl. Grafika 2, Yogyakarta,
Indonesia
Email : [email protected]
† Dept. of Electrical Engineering, Fac. of Engineering, Gadjah Mada University, Jl. Grafika 2, Yogyakarta,
Indonesia
Email : [email protected]
Abstract – Turbo Code was proposed in 1993 by Berrou,
Glavieux, and Thitimajashima, who reported excellent
coding gain results (Berrou, 1993), approaching
Shannonian predictions. The turbo code uses two
Recursive Systematic Convolutional (RSC) encoders
separated by an interleaver. There are two main
decoding techniques which one of them is Soft Output
Viterbi Algorithm (SOVA). SOVA is a modified Viterbi
Algorithm making it is able to use a-priori information
between SOVA decoders. Reliability estimation, loglikelihood algebra, and soft channel outputs are the
fundamentals for SOVA decoding. The iterative SOVA
decoding algorithm is used to increase the performance
of turbo code. Field Programmable Gate Array (FPGA) is
a reprogrammable digital chip which is considered as
very fast and easy to modify chip making it suitable for
prototyping products. Based on above considerations,
FPGA (EPF10K10LC84) is choosen to implement the
Turbo Code.
There are 10 units to be implemented; encoder, 8
units composing a decoder, and control unit. Both the
design process and the implementation use MAX+plus II
software.
The result shows that the simulation works well,
although it cannot to be implemented in a single chip.
The circuit needs 1553 LEs which is much bigger than
the available LEs on EPF10K10LC84. The
implementation results show it is able to correct a
maximum of 5 errors bits.
Viterbi Algorithm (SOVA). SOVA is a modified Viterbi
Algorithm making it is able to use a-priori information
between SOVA decoders. Reliability estimation, loglikelihood algebra, and soft channel outputs are the
fundamentals for SOVA decoding. The iterative SOVA
decoding algorithm is used to increase the performance of
turbo code.
FPGA is a reprogrammable chip. A design in FPGA
can be automatically converted from gate level into layout
structure by place and route software.
This research is our preliminary research on
implementation Turbo Code in FPGA.
II. BASIC THEORY
Recursive Systematic Convolutional (RSC) encoder
RSC is developed from Convolution encoder. RSC
.
, and
encoder written as
are code generators, G1 is feedback polynomial
generator and G2 is forward polynomial generator. Figure 1
show us Recursive Systematic Convolutional encoder block
diagram with constraint length (K) = 3 and ak is obtained
from equation (1).
(1)
Where g1 = g1i if uk = dk and g2 = g2i if vk = dk
Keywords - Turbo Code, RSC Encoder, SOVA Decoder,
Soft Output, FPGA (Field Programmable Gate Array)
I. INTRODUCTION
Error Correcting Control is very important in modern
communication systems which one of them is Turbo Code.
Turbo Code was proposed in 1993 by Berrou, Glavieux, and
Thitimajashima, who reported excellent coding gain results
(Berrou, 1993), approaching Shannonian predictions. The
turbo code uses two Recursive Systematic Convolutional
(RSC) encoders separated by an interleaver. There are two
main decoding techniques which one of them is Soft Output
Figure 1. Recursive Systematic Convolutional Encoder
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350 Proceedings of CITEE, August 4, 2009 52
SOVA decoder
Turbo decoder is a modified Viterbi algorithm. The
aim of this modification is to make Viterbi algorithm
provides soft output values. These values are given to one of
SOVA decoders as a-priori information in order to
increasing the decoder performance. SISO algorithm
receives a-priori information and provides a-posteriori as
output (Figure 2).
information and error series into serial output. There are two
main parts to be implemented: Information series and error
series generator. The truth table of information series
generator is given by Table 1.
Table 1. Truth Table of Information Series Generator
Figure 2. SISO Decoder
There are two modifications in Viterbi algorithm to
make SOVA decoders able to provide soft output:
1. Defined ML path, SOVA uses former matrix path
information and a-priori information to define the ML
path or chosen path.
2. Update process, SOVA must be able to provide soft
output (a-posteriori) which is used as a-priori input
for another SOVA.
Based on the truth table in Table 1, the VHDL
program of Information series generator is shown in Figure
3.
SOVA decoder use three kinds of information which is
different from Viterbi:
1. A-priori
A-priori information is encoded bit probability which
is known before encoding process. Obtained from
another decoder’s output and called as intrinsic
information.
2. Extrinsic
Extrinsic information is encoded bit probability which
is obtained from a-posteriori by deleting the effect of
and a-priori
received systematic information
.
information
3. A-posteriori
A-posteriori information is state series probability
which is obtained based on soft channel output
.
information and a-priori information
III. EXPERIMENTAL RESULTS
Turbo Encoder and Decoder designed using
MAX+plus II software. The designs are implemented using
VHDL hardware programming language. Turbo Codes
System is divided into 5 main units: parallel into serial
converter (information series and error series generator),
encoder, decoder, serial into parallel converter, and control
unit.
Parallel into Serial Converter
Encoding and Decoding in Turbo codes are processed
in serial way but information and error series are given in
parallel, so we need parallel into serial converter to convert
ISSN: 2085-6350 Figure 3. VHDL Program for Information Series Generator
The truth table of error series generator is given by
Table 2.
Conference on Information Technology and Electrical Engineering (CITEE) 53 Proceedings of CITEE, August 4, 2009
Table 2. Truth Table of Error Series Generator
Encoder
Figure 5 shows the block diagram of Turbo encoder.
Figure 5. Turbo Encoder Block Diagram
Turbo encoder is obtained from two registers (Flip
Flop D) which is designed in VHDL programming language
as shown in Figure 6.
Based on the truth table in Table 2, the VHDL
program of Information series generator is shown in Figure
4.
Figure 6. VHDL Program for FFD
Decoder
SOVA decoders can be implemented by dividing it
into 8 sub units: Branch Metric Generator (BMG), Add
Compare and Select Unit (ACSU), Delta Path Metric
(DeltaPM), Maximum State Select Unit (MSSU), Track
Back Unit (TBU), Path Equivalence Detector (PED),
Reliability Measurement Unit (RMU), and A- posteriori
Measuring Unit.
Serial into Parallel Converter
Serial into Parallel circuit is given by Figure 7.
Figure 4. VHDL Program for Error Series Generator
Figure 7. Serial into Parallel Converter Circuit
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350 Proceedings of CITEE, August 4, 2009 54
A-posteriori Circuit and Decoded Bit
Figure 8 shows us the circuit which is provide aposteriore result and decoded bit.
and encoded bit are determined at TBU. TBU needs the
biggest LE, 558 LE or 96.88% of EPF10K10LC84’s
maximum capacity.
IV. CONCLUSIONS
The Turbo Codes implementation results show it is
able to correct random error up to 5 bits. SOVA decoders
can be implemented by dividing it into 8 sub units: Branch
Metric Generator (BMG), Add Compare and Select Unit
(ACSU), Delta Path Metric (DeltaPM), Maximum State
Select Unit (MSSU), Track Back Unit (TBU), Path
Equivalence Detector (PED), Reliability Measurement Unit
(RMU), and A-posteriori Measuring Unit.
Chosen metric path and encoded bit are determined at
TBU. TBU needs the biggest LE, 558 LE or 96.88% of
EPF10K10LC84’s maximum capacity.
Figure 8. A-posteriore Circuit
Control Unit
Figure 9 shows us the circuit which is provide control
signal to control some units and components.
Figure 9. Control Signal Generator Circuit
The Turbo Codes implementation results show it is
able to correct random error up to 5 bits. Chosen metric path
ISSN: 2085-6350 REFERENCES
[1] Altera Corporation, January 2003, “FLEX 10K
Embedded Programmable Logic Device Family”, Data
Sheet version 2.2, Altera Corporation.
[2] Altera Corporation, January 2003, “FLEX 10K
Embedded Programmable Logic Device Family”, Data
Sheet version 4.2, Altera Corporation.
[3]Altera Corporatio, July 2002, ”Turbo Encoder/Decoder”,
Megacore Function User Guide Version 1.1.2 Rev. 1,
Altera Corporation.
[4] Berrou, C., 1993, Turbo Codes: Some Simple Ideas for
Efficient Communications.
[5]Hagenauer, J., Hoeher, P., 1989, “A Viterbi with SoftDecision Outputs and its Applications”, German
Aerospace research Establishement, Institute for
Communiaction Technology, West-Germany.
[6]Hanzo,L., Liew, T.H., Yeap, B.L., 2002, “Turbo Coding,
Turbo Equalisation and Space-Time Coding”,
Department of Electronics and Computer Science,
University of Southampton, UK.
[7]Huang, F., May 29, 1997, “Evaluation of Soft Output
Decoding for Turbo Codes”, Master's thesis, Virginia
Polytechnic Institute and State University, Blacksburg,
Virginia.
[8]Kusumawardani S.S., 2001, “Implementasi Sandi BCH
(15,5) dengan FPGAXC4013”, EE UGM, Yogyakarta.
[9]Lin,S. dan Castello, 1983, “Error Control Coding:
Fundamental
and
Apliccation”,
Prentice-Hall
International Inc, Englewood Cliff, New Jersey.
Conference on Information Technology and Electrical Engineering (CITEE) Proceedings of CITEE, August 4, 2009
55
BER Performance Analysis of PAM and PSM for
UWB Communication
Risanuri Hidayat
Electrical Engineering Department
Faculty of Engineering Gadjah Mada University
Yogyakarta 55281, Indonesia
[email protected]
Abstract— This paper proposes BER analysis of Impulse
Radio Ultra Wideband (IR-UWB) pulse modulations. A
Gaussian monocycle as UWB pulse is modulated using Pulse
Amplitude Modulation (PAM) and Pulse Shape Modulation
(PSM). The channel model is generated from a modified S-V
model. Bit-error rate (BER) is measured over several of bit
rates. The result shows that PAM gives better performance
than PSM in bit rates and SNR. Moreover, as standard of
speed has been given for UWB, both modulations is
appropriate with high bit rates in LOS channel.
where Tc = 1 / f c is the width of the pulse. The values A1 and
A2 are amplitudes. The Federal Communications
Commission (FCC) has recently approved the deployment of
UWB on an unlicensed basis in the 3.1–10.6 GHz band
subject to a modified version of Part 15.209 rules as in Fig. 1
[1]. The essence of this regulation is to limit the power
spectral density (PSD) measured in a 1–MHz bandwidth at
the output of an isotropic transmit antenna.
Keywords— IR-UWB, S-V Channel Model, LOS NLOS,
PAM, PSM
I.
INTRODUCTION
Since FCC released UWB frequency band [1], UWB
performance for communication has been studied [2]-[13],
including binary modulations for impulse radio include pulse
amplitude modulation (PAM) [3] and pulse shape
modulation (PSM) [4]. BER performance for impulse radio
ultra wideband (IR-UWB) systems in multipath propagation
channels has also been introduced [5][6][7][8][9]. In order to
realize actual condition, a statistical model is established for
the ultra-wide bandwidth in a simulation [10][11], and the
channel model using a Saleh-Valenzuela (S-V) [12] that
showed multipath arriving in clusters has been modified for
UWB signal application [13].
This paper presents the analysis of modulations of UWB
through S-V channel that has been modified for UWB
communications [13] and AWGN noise especially in bit
rates and BER by a simulation. PAM and PSM are
implemented for this purpose. Data sent by the transceiver
are converted to modulated UWB pulse and are then
received by the receiver. At the receiver the pulse is then
converted to data again. The comparison between the data
bits sent by transmitter and the data bits received by receiver
represents the BER values.
II.
UWB PULSE AND ITS MODULATIONS
Some references have represented several formulas of
Gaussian pulse [14] as an IR-UWB. A monocycle pulse is
the first derivative of Gaussian pulse. The Gaussian pulse
has formula (1), while the monocycle is given in (2).
x ( t ) = A1 .e
⎛π t ⎞
−2⎜
⎟
⎝ Tc ⎠
2
x
⎟
− 2 ⎜⎜
dx ( t )
T ⎟
(t ) =
= A 2 t .e ⎝ c ⎠
dt
A number of modulation schemes may be used with IRUWB systems. The potential modulation schemes include
both orthogonal and antipodal schemes. Some common
methods of modulation are to invert the called as pulse
amplitude modulation (PAM) [3] and pulse shape
modulation (PSM) [4] [15], which requires special pulse
shapes to be generated which are orthogonal to each other.
A. Pulse Amplitude Modulation
PAM is implemented by binary pulse modulation, which
is presented using two antipodal Gaussian pulses. The
transmitted binary baseband pulse that is modulated
information signal (t) is expressed as [3]
x(t ) = d j .wtr (t )
(3)
where wtr(t) represents the UWB pulse waveform, j
represents the bit transmitted that d j = −1, j = 0 and
d j = +1, j = 1
(1)
⎛ πt ⎞
(1 )
Figure 1. UWB spectral mask and FCC Part 15 limits
2
(2)
Fig 2 (a) shows the pulse amplitude modulation of the
Gaussian monocycle pulse waveform expressed in (2) with
Tc = 2.22x10-1 ns. Fig. 2 (b) shows the normalized Power
Spectral Density (PSD) of the derivative of pulse. When a
pulse is transmitted, due to the derivative characteristics of
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
56
Proceedings of CITEE, August 4, 2009
the antenna, the output of the transmitter antenna can be
modelled by the first derivative of the pulse [3]. The
derivative can be achieved by the basics of differential
calculus
x (1) ( t ) =
lim
dt → 0
x ( t + dt ) − x ( t )
dt
(4)
PSD of the deterministic power signal, w(t), is
P( f ) =
lim
T →∞
⎛ X(f)2 ⎞
⎜
⎟
⎜
⎟
T
⎝
⎠
(a) PSM pulse shapes
(5)
where T is the pulse spacing interval. X(f) is the Fourier
transform of the pulse, i.e. x(t). P(f) has units of watts per
hertz. When X(f) is attained, the peak emission frequency,
i.e. fM, can be found as the frequency at the maximum value
of |X(f)|. The normalized PSD is used to comply with the
FCC spectral mask of the pulse that is transmitted by
antenna [1][16][17]. The normalized PSD can be defined as
follows
(b) Frequency domain
P( f ) =
X(f)
2
X ( fM )
2
(6)
Fig. 2 (b) shows that the pulses have normalized PSD in
3.1-10.6 GHz frequency band that fulfil the FCC rules of
UWB communication as shown in Fig. 1.
B. Pulse Shape Modulation
Pulse shape modulation (PSM) uses different, orthogonal
waveforms to represent bit ‘0’ and ‘1’. The transmitted pulse
can be represented as
Figure 3. Pulse waveforms used for PSM modulation and its frequency
domain
x(t ) = (1 − d j ) wtr( 0) (t ) + d j wtr(1) (t )
(7)
where dj is defined as in (8) and wtr(0) (t ) and wtr(1) (t ) represent
two different pulse waveforms.
Fig. 3 shows the pair of PSM pulses and it’s PSD. The
pulse waveforms apply the Gaussian monocycle pulse
defined as in (2) with Tc = 2.22x10-1 ns. The second
waveform is the derivative of (2). It means that the pulse
shapes are the first and the second derivatives of the
Gaussian pulse. These two waveforms are orthogonal
according to the definition given by the cross correlation of
the two waveforms,
+∞
ρ c (τ ) | t =0 =
∫w
tr
(t ).wtr (t + τ )dt = 0
(8)
−∞
Fig. 3 (b) shows that both of the PSM pulses have the
normalized PSD that comply with the FCC spectral mask.
(a) BPAM pulse shapes
III. CHANNEL MODEL
Since UWB waveforms can be up to 7.5 GHz wide, for
example, any paths separated by more than about 133 psec.
(equivalent to 4 cm path length difference) can be
individually resolved at the receiver [18]. The realistic
channel for IEEE 802.15 study group 3a has been developed
by Saleh-Valenzuela (S-V model) and proposed for the real
indoor channel model, where the clusters and rays are multipath components [13]. This multi-path channel can be
expressed as
(b) Frequency domain
Figure 2. PAM pulse shapes for ‘1’ and ‘0’ bits and its frequency domain
ISSN: 2085-6350
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
L
57
K
h(t ) = ∑∑ α k ,l δ (t − Tl − τ k ,l )
l =0 k =0
(9)
where α k,l
is the multipath gain coefficient, Tl is the
delay of the lth cluster, and τ k,l is the delay of the kth
multipath component relative to the lth cluster arrival time
( Tl ).
As suggested in [13], α k,l = p k,l β k,l is adopted,
where p k,l is equally likely to take on the values of +/-1,
and β k,l is the lognormal fading term, due to the
simplicity of the real channel coefficients, and to avoid the
ambiguity of phase for an UWB waveform.
The proposed channel model uses the definitions as
previously described in the S-V model, and is repeated here
for completeness. Tl is the arrival time of the first path of the
l-th cluster; τk,l is the delay of the k-the path within the l-th
cluster relative to the first path arrival time, Tl; Λ is cluster
arrival rate; and λ = ray arrival rate, i.e., the arrival rate of
path within each cluster.
Therefore, τ 0l = Tl . The distribution of cluster arrival
time and the ray arrival time are given by
p(Tl Tl −1 ) = Λexp ⎡⎣−Λ(Tl −Tl −1 ) ⎤⎦ , l > 0
(
)
(
)
p τk,l τ(k−1),l = λ exp ⎡⎣−λ τk,l −τ(k−1),l ⎤⎦ , k > 0
(10)
The channel coefficients are defined as α k ,l = p k ,l β k ,l ,
β k ,l is obtained by this expression
20 log10( β k ,l ) ∝ Normal( µ k ,l , σ 2 )
or
IV. SIMULATION METHOD
Fig. 4 shows the block diagram of the transmitter and
receiver communication. Transmitter converts the data bits
to UWB pulses and the pulse are transmitted through the
antenna. There is a channel model between the transmitter
and the receiver. At receiver, the pulses are received by
antenna and LNA (Low Noise Amplifier) becomes uf(t).
Finally, the pulses are detected using the template pulse. The
Gaussian monocycle pulse expressed in (2) with Tc =
2.22x10-1 ns is implemented. PAM and PSM modulations
are implemented using (3) and (7), respectively.
A channel is available between transceiver and receiver.
An UWB channel model is derived from the modified SalehValenzuela model suitable for home environment with both
LOS and NLOS as described previously. There are 5 key
parameters that define the model: Λ = cluster arrival rate; λ
= ray arrival rate, i.e., the arrival rate of path within each
cluster; Γ = cluster decay factor; γ = ray decay factor; and σ
= standard deviation of lognormal fading term (dB). All
parameters are taken into account for channel modeling as
formulated in (10)-(13).
This paper uses one channel model generated from S-V
model using the parameters as shown in Table I. The
equation in (9)-(13) and Table I derived from and based on
Intel measurements [13] are used to generate a channel
model. The generated channel model is shown in Fig 10.
It should be noted that in an ideal environment the
received pulse shape is the second derivative of the signal
transmitted due to differentiator effect in transmitter and
receiver antennas. However, the amplitude of correlation
result between the first and the third derivatives of the
Gaussian pulse is 80% of the autocorrelation of the received
waveform [3]. This means that using the transmitted
waveform as the template, the performance of the correlator
is reduced by less than 1 dB. Therefore, the receiver still can
use the same template waveform to demodulate the signals.
(11)
β k ,l = 10 n / 20
where
n ∝ Normal( µ l , σ 2 )
[ ]
E β k2,l = Ω 0 e −Tl / Γ e
(12)
−τ k ,l / γ
Tl is the excess delay of bin l and Ω 0 is the mean power of
Figure 4. UWB pulse communication
the first path of the first cluster, and p k ,l is +/-1. Then µl is
TABLE I.
given by
Model Parameters
Λ (1/nsec)
λ (1/nsec)
Ѓ (nsec)
γ (nsec)
σ (dB)
µl =
10 ln( Ω 0 ) − 10Tl / Γ − 10τ k ,l / γ
ln(10 )
−
σ 2 ln(10)
20
(13)
Conference on Information Technology and Electrical Engineering (CITEE)
CHANNEL MODEL PARAMETERS.
LOS
1/60
1/0.5
16
1.6
4.8
NLOS
1/11
1/0.35
16
8.5
4.8
ISSN: 2085-6350
58
Proceedings of CITEE, August 4, 2009
it is much thicker. The bit rates drops drastically in the
NLOS channel model. Note that in order to draw the line in
Fig. 7, the value of log (10-6) is given instead of BER = 0,
since log zero is infinity.
(a) LOS
(a) PAM over LOS channel
(b) NLOS
Figure 5. The generated channel model
V. SIMULATION RESULTS
The implementation based on the block diagram in Fig. 4
shows the UWB pulse communication. The transceiver
converts bits into UWB pulses, then the UWB pulses are
reconverted the pulse back into bits by receiver. The
comparison between the bits sent by transceiver and received
by receiver gives BER value.
(b) PSM over LOS channel
The Gaussian monocycle pulse defined in (2) is
implemented with Tc = 2.22x10-1 ns and PAM and PSM
modulations are applied as in (3) and (7). The channel is
considered to analyze the bit rates of each pulse over the
channel. The channel models, both LOS and NLOS using
parameter in Table I are implemented as shown in Fig. 5.
Wireless communication systems that use frequency
band between 3.1-10.6 GHz. GSM/UMTS uses 0.9-1.8GHz
is rare, while WLAN uses 2.5 GHz. Only 802.11a system
uses frequency band 5 GHz inside this UWB frequency
standard. Therefore, some AWGN are included in the
simulation.
Our experiments apply 104 bits sent from the transceiver
to receiver, and later being analyzed. The bits are sent in
Mbps bit rates. The various bit rates and its BER can be seen
in Fig. 7. Fig. 6 shows the result of the BER of PAM and
PSM pulse modulation of IR_UWB through the LOS and
NLOS channel models. Fig. 6 shows that the influence of
inter multipaths occurs in higher bit rates. The LOS channel
model in Fig. 5 (a) has the last significant paths that cross a
10 dB threshold (NP10dB) at 0.5x10-7 sec and it has only 5
multipaths. It gives low BER until hundreds Mbps.
Meanwhile, the NLOS channel has many multipaths that
influence the pulses, so that it can not be sent in high bit
rates. The NLOS channel model in Fig. 5 (b) has the last
significant paths that cross a 10 dB threshold (NP10dB) at
around 1.0x10-7 sec and it has around 33 multipaths, so that
ISSN: 2085-6350
(c) PAM over NLOS channel
(d) PSM over NLOS channel
Figure 6.
BER of the UWB pulses through the LOS and NLOS channel
models
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
Fig. 6 (a) shows data sent using PAM over the LOS
channel model. The result shows that SNR 4 and 8 dB
AWGN give high BER. SNR 12 dB gives no error (or error
lower than 10-4) until 800 Mbps. BER increases drastically
after 800 Mbps.
Fig. 6 (b) shows data sent using PSM over the LOS
channel model. SNR 16 and 20 dB AWGN give high BER.
SNR 24 dB gives error in 10-3 level until 800 Mbps and BER
increases afterwards. SNR 28 dB has no error (below 10-4)
until 800 Mbps, and the value of BER increases drastically
afterwards. Therefore, this modulation needs higher SNR
than the other modulations.
Different from the LOS channel that has much higher bit
rates, the NLOS channel gives lower bit rates for the
communication. In Fig. 6 (c), data are sent using PAM over
NLOS channel model. It has high BER until SNR 12 dB.
SNR 16 dB has no error (or error lower than 10-4) within 10
Mbps, gives BER in 10-3 level for bit rates below 30 Mbps
and BER increases drastically after 30 Mbps. SNR 20 dB
has no error (or error lower than 10-4) until 30 Mbps, but it
has higher BER afterwards.
In Fig. 6 (d), data are sent using PSM over the NLOS
channel model. It has high BER until SNR equals to 28 dB.
SNR 32 dB gives BER in 10-4 level for 10 Mbps and BER in
10-3 level until 30 Mbps. It has also high BER for above 30
Mbps.
Comparing among the modulations, the result shows that
for the same channel, PAM is more robust of AWGN than
PSM, both in LOS and NLOS channels. From the results, it
can be said that UWB pulse communication is appropriate in
LOS channel, i.e., has high bit rates (hundreds Mbps).
Nevertheless, this communication can be used until several
Mbps in NLOS channel model.
TABLE II.
LOS
PAM
PSM
NLOS
PAM
PSM
ESTIMATE OF HIGHEST BIT RATES (HBR)
Bit rates (Mbps)
SNR 16 dB
SNR 12 dB
800
800
Bit rates (Mbps)
SNR 32 dB
SNR 20 dB
SNR 16 dB
30
30
30
30
SNR 28 dB
800
800
Table II shows estimation of highest bit rates without
error (or error less than 10-4 level is tolerable). If the less
BER is desired, then lower bit rate can be implemented.
However, in high SNR environment, the high bit rate can
still be implemented. It shows that PAM modulation over
LOS channel gives the best performance. By looking at the
UWB reference standard and compare to the others as shown
in Table III [19], it seems that this communication model is
preferable for the LOS channel, but PSM needs high SNR
environment.
Based on Fig. 6(a)-(d), some samples of bit rate are
chosen to take a closer look at each modulation over LOS
and NLOS channels. Based on Table III, 400 Mbps is chosen
for all modulations for the LOS channel. However, since the
maximum bit rates for NLOS channel is only 30 Mbps, this
rate is chosen.
59
There are 104 bits sent from the transceiver to receiver in
Fig. 9. The figure shows the BER performance of the UWB
pulses through the LOS and NLOS channel models in the
chosen bit rates. Fig 13 shows that PAM is more robust both
in LOS and NLOS channels.
TABLE III.
COMPARISON OF UWB BIT RATE WITH OTHER
STANDARDS.
Speed (Mbps)
480
200
110
90
54
20
11
10
1
Standard
UWB, USB 2.0
UWB (4 m minimum)
UWB (10 m minimum)
Fast Ethernet
802.11a
802.11g
802.11b
Ethernet
Bluetooth
When 400 Mbps is chosen for LOS channel model, it
means that the pulse is sent in 2.5ns periodic time. PAM has
BER in 10-3 level for SNR 8 dB, and low BER (in 10-4
level or lower) afterwards. Fig. 7 shows that data
communication using PAM that needs BER 10-5 or less can
be applied in SNR 12 dB or more.
PPM has BER in 10-2 level within SNR 10 dB, and in
10-4 level for SNR 10 dB. It can be estimated that using
PPM, BER 10-5 or less can be applied in SNR 14 dB or
more.
PSM has BER in 10-2 level within SNR 22 dB, BER in
10-3 level for SNR 24 dB, and BER in 10-4 level for SNR
26 dB. It can be estimated that using PPM, BER 10-5 or less
can be applied in SNR 30 dB or more.
When 30 Mbps is chosen for NLOS channel model, it
means that the pulse is sent in 0.33 ns periodic time. PAM
has BER in 10-2 level for SNR 12 dB, BER in 10-4 for SNR
14 dB, and low BER (lower than 10-4 level) afterwards. Fig.
7 shows that data communication using PAM that needs
BER 10-5 or less can be applied in SNR 16 dB or more.
PSM has BER in 10-2 level within SNR 28 dB, BER in
10-3 level for SNR 30 dB, and BER in 10-4 level for SNR
32 dB. It can be estimated that using PPM, BER 10-5 or less
can be applied in SNR 34 dB or more.
VI.
CONCLUSION
The BER analysis of the Impulse Radio Ultra Wide Band
pulse communication over modified S-V channel model is
presented. The pulse uses Gaussian monocycle pulse. The
channel model uses a modified S-V model introduced by
Intel. PAM and PSM are applied for modulation, and BER is
measured over several bit rates. The result shows that all
modulation are appropriate for both LOS and NLOS
channel, but PAM gives the best performance in bit rates and
SNR. Moreover, as standard of speed has been given for
UWB, the communication is appropriate with high bit rates
in LOS channel.
ACKNOWLEDGMENT
The authors would like to thank the anonymous
reviewers for their constructive comments and insights that
help to improve this paper.
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
60
Proceedings of CITEE, August 4, 2009
[5]
[6]
[7]
[8]
[9]
[10]
(a)
[11]
[12]
[13]
[14]
[15]
[16]
(b)
Figure 7. BER performances over channels
[17]
REFERENCES
[1]
[2]
[3]
[4]
[1] FCC 04-48 “First Report and Order in the Matter of Revision of
Part 15 of the Commission’s Rules Regarding Ultra-wideband
Transmission Systems.” Feb.14, 2002.
[2] R. A. Scholtz, “Multiple access with time-hopping impulse
modulation,” Proc. MILCOM ’93, vol. 2, pp. 447-450, 1993.
[3] I. Oppermann, M. Hamalainen, and J. Iinatti, “UWB Theory and
Applications”,John Wiley & Sons, West Sussex, 2004.
[4] L. B. Michael, M. Ghavami, and R. Kohno, “Multiple pulse
generator for ultra-wideband communication using Hermite
polynomial based orthogonal pulses,” in IEEE Conf. Ultra Wideband
Systems and Technologies, Dig. Papers, Baltimore, MD, 2002, pp.
47–51.
ISSN: 2085-6350
[18]
[19]
[5] W. Xu, R. Yao, Z. Guo, W. Zhu, and Z. Zhou, “A power efficient
M-ary orthogonal pulse polarity modulation for TH-UWB system
using modified OVSF codes”, in Proc. IEEE Global
Telecommunications Conf. (GLOBECOM), San Francisco, CA, Dec.
1–5, 2003, vol. 1, pp. 436–440.
[6] S.-C. Lin and T.-D. Chiueh, “Performance analysis of impulse
radio under timing jitter using M-ary bipolar pulse waveform and
position modulation”, in IEEE Conf. Ultra Wideband Systems and
Technologies, Dig. Papers, Reston, VA, Nov. 16–19, 2003, pp. 121–
125.
[7] D. J. Clabaugh, M. A. Temple, R. A. Raines, and C. M. Canadeo,
“UWB multiple access performance using time hopped pulse position
modulation with biorthogonal signaling”, in IEEE Conf. Ultra
Wideband Systems and Technologies, Dig. Papers, Reston, VA, Nov.
16–19, 2003, pp. 330–333.
[8] L. Ge, G. Yue, and S. Affes, "On the BER Performance of PulsePosition Modulation UWB Radio in Multipath Channels", Proc. IEEE
Conference on Ultra Wideband Systems and Technologies, pp. 231234, May 2002.
[9] G. Yue, L. Ge, and S. Li, “Performance of UWB time-hopping
spread-spectrum impulse radio in multipath environments”, Proc.
IEEE Semiannual Vehicular Technology Conference (VTC '03), vol.
3, pp. 1644–1648, April, 2003.
[10] J. Karedal, S. Wyne, P. Almers, F. Tufvesson and A.F. Molisch,
“Statistical Analysis of the UWB Channel in an Industrial
Environment”, Proc. IEEE Vehicular Technology Conference (VTC
2004), pp. 81-85, Sept. 2004.
[11] D. Cassioli, M. Win and A. Molisch, "The Ultra-Wide
Bandwidth Indoor Channel: From Statistical Model to Simulations",
IEEE Journal on Selected Areas in Communications, Vol. 20, No. 6,
pp. 1247-1257, 2002.
[12] A. M. Saleh, R. A. Valenzuela, “A statistical model for indoor
multipath propagation”, IEEE Journal on Selected Areas in
Communications, SAC-5 no.2, pp.128- 137, 1987.
[13] J. R. Foerster, Q. Li, “UWB Channel Modeling Contribution
from Intel”, IEEE P802.15 Working Group for Wireless Personal
Area
Networks
(WPANs),
June
2002,
http://www.ieee802.org/15/pub/2002/Jul02/ 02279r0P802-15_SG3aChannel-Model-Cont-Intel.doc
[14] X. Chen, and S. Kiaei, “Monocycle shapes for ultra wideband
system”, Proc. IEEE International Symposium on Circuits and
Systems, vol. 1, pp. 597-600, May 2002.
[15] R. S. Dilmaghani, M. Ghavami, B. Allen, and H. Aghvami,
“Novel UWB pulse shaping using prolate spheroidal wave
functions”, in Proc. IEEE Personal, Indoor and Mobile Radio
Communications Conf. (PIMRC), Beijing, China, Sep. 7–10, 2003,
vol. 1, pp. 602–606.
[16] G. Breed, “A Summary of FCC Rules for Ultra Wideband
Communications,”2005,http://www.highfrequencyelectronics.com/Ar
chives/ Jan05/
[17] S. Hongsan, P. Orlik, A.M. Haimovich, L. J. Jr. Cimini, and Z.
Jinyun, “On the Spectral and Power Requirements for UltraWideband Transmission”, Proc. IEEE on Communications, vol. 1,
pp. 738 – 742, May 2003.
[18] A.F. Molisch, J.R. Foerster, M. Pendergrass, “Channel models
for Ultra wideband personal area networks”, IEEE Wireless
Communications, vol. 10, Issue 6, pp.14-21, 2003.
[19] M. Ghavami, L. B. Michael, R. Kohno, “Ultra Wideband
Signals and Systems in Communication Engineering”, John Wiley &
Sons, West Sussex, 2004.
4ll
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
61
Hardware Model Implementation of a
Configurable QAM Mapper-Demapper for an
Adaptive Modulation OFDM System
Budi Setiyanto, Astria Nur Irfansyah, and Risanuri Hidayat
Electrical Engineering Department, Faculty of Engineering, Gadjah Mada University
Jl. Grafika 2 Yogyakarta, 55281, Indonesia
[email protected]
Abstract—This research implemented a configurable
4/16/64/256 quadrature amplitude modulation (QAM) mapperdemapper as a part of an adaptive modulation orthogonal
frequency division multiplexing (OFDM) hardware model. The
model is developed as a learning-equipment. Therefore, all
electronic components are selected from the general purpose
types, and performance quality is beyond the target. Mapper
and demapper are a sequential logic circuit controlled by a
synchronized cyclic finite state machine (FSM). The standalone latency generated by mapper and demapper are as long
as 8.5 and 0.5 bits duration, respectively. But, this 0.5 bit
demapper latency will not appear when this mapper-demapper
couple is integrated to the other parts in the QAM or OFDM
model. This mapper-demapper can be configured to form a 4,
16, 64, or 256 orders QAM system, and capable for supporting
a bit rate up to about 4 Mbps.
Keywords—adaptive modulation, OFDM, QAM
If one bit duration is Tb, then both I(t) and Q(t) have
constant value during every T = 2NTb duration.
B. Conceptual OFDM
At the K sub-channels OFDM system transmitter-side,
the original bit stream, d(t), is split to K sub-streams. Each
sub-stream is QAM modulated. Thus, conceptually, the
OFDM transmitter can be visualized in Figure 1.
I. INTRODUCTION
In any communication system, the transmission power
and frequency bandwidth are strictly limited, while the bitrate, noise-immunity, and fading-robustness should be
optimized. OFDM employing QAM is one of suitable
techniques for compromising such resource constrain in one
side and performance objective in other side [1, 4]. It is (and
will be) widely applied in recent (and future) wired and
wireless technologies [2 – 6]. Adaptive modulation is a
scheme intended to optimize the performance.
A configurable 4/16/64/256-QAM hardware model has
been implemented in this research. It is a part of an adaptive
modulation OFDM hardware model. The model is developed
as a learning-equipment. Therefore, all electronic
components are selected from the general purpose types, and
performance quality is beyond the target.
II. REVIEW OF THEORETHICAL BACKGROUND
A. M-QAM
At the M-QAM (M = 22N, N = 1, 2, 3, …) transmitterside, the original bit stream, b(t), is broken-down to 2N-by2N bits segments. Each 2N-bit segment is further split to two
N-bit length groups. Alternate splitting is more suggested
than sequential. In such splitting scheme, the first and second
groups are built of N even-numbered bits and the N oddnumbered bits, respectively. Those two groups are converted
to in-phase and quadrature baseband analog signals notated
as I(t) and Q(t), respectively. Using M(t) = I(t) – jQ(t), the
QAM signal can be expressed as
{
(
s(t ) = Re M (t )exp j 2πf t
c
)}
Figure 1. Conceptual OFDM transmitter visualization
According to equation (1), this OFDM signal can be
expressed as
(
)
K
s' (t ) = Re ∑ M (t )exp j 2πf t = Re(s(t ))
k
k
k =1
(2)
In equation (2), each Mk(t) has constant value during the
symbol duration of TM = 2KNTb. To preserve the
orthogonality, the sub-carrier frequency, fk for all k, must be
an integer multiply of fM = 1/TM. Figure 1 is not feasible to
be realized (implemented) directly.
C. Feasibly-Realized OFDM
Let fk = fc + kfM. By choosing fc a sufficiently high
integer multiply of fM, the summation index (k) in equation
(2) must not be started from a positive integer. By changing
the summation index, the complex-valued signal in equation
(2), can be expressed as
(
K −1
s(t ) = ∑ M (t )exp j 2πf t
k
k
k =0
)
(3)
Sampling s(t) as many as K times during TM, or with
sampling interval of Ts = TM/K, gives
( )
K −1
s( nT ) = ∑ M nT exp( j 2πkn / K )
s
k s
k =0
(4)
(1)
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
62
Proceedings of CITEE, August 4, 2009
For any k, Mk(nTs) has constant value, that is Mk(nTs) =
M(k) for all n in this k. Normalizing Ts = 1 and replacing
Mk(nTs) with M(k), yield
K −1
s( n) = ∑ M (k )exp( j 2πkn / K ) = IDFT{M (k )}
k =0
(5)
B. Principles of the OFDM Hardware
According to the previous feasible-realized OFDM
review, principle of OFDM transmitter can be modelled as
shown in Figure 4, and the associated receiver in Figure 5.
Equation (5) shows that the complex-valued OFDM
signal is the inverse discrete Fourier transform (IDFT) of the
M(k) = Mk(t) for mTM < t < (m + 1)TM. Using inverse fast
Fourier transform (IFFT) processor, equation (5) can be
implemented efficiently. Therefore, transmitting the
complex-valued OFDM signal equation (5) is more feasible
rather than transmitting the real-valued one equation (2).
Figure 4. Principle of OFDM transmitter
D. Adaptive Modulation OFDM
Bit error rate (BER) increases as the Eb/No decreases.
Effort which can be done to prevent the Eb/No does not fall
below the desired level without increasing the power and
still occupying the overall available bandwidth is decreasing
the operating QAM order. OFDM in which its QAM order is
changeable like this is called adaptive modulation OFDM.
III. PRINCIPLES FOR IMPLEMENTATION
Figure 5. Principle of OFDM receiver
All of the implemented models are constructed as a
trainer, and all of the digital integrated circuit components
are chosen from 74LSxx TTL series. All of the design,
simulation, and lay-out works are aided by OrCAD Release
9.1 software package.
Figure 4 and 5 are generic form. For real applications,
additional sections may be inserted to improve the system
performance [4 – 6].
IV. IMPLEMENTATION OF MAPPING TECHNIQUE
A. Principles of the QAM Hardware
According to the previous M-QAM theoretical review,
principle of QAM transmitter can be modelled as shown in
Figure 2, and the corresponding receiver in Figure 3.
Figure 2. Principle of QAM transmitter
Previous work has been implemented a fix 16-QAM
mapper for four channels OFDM [7].
The QAM mapper implemented in this research is a
configurable (selectable) 4, 16, 64, or 256 orders. Block
diagram of this mapper is shown in Figure 6. In this figure,
notation bIN and m are used rather than b(k) and m(k),
respectively. The QAM symbol, m, takes a complex form as
m = I + jQ, where I and Q represent in-phase and quadrature
components, respectively. This diagram shows two parts: the
signal flow section and control circuit section at the upper
and lower parts, respectively. The signal flow part consists
of one bit-by-bit serial-to-parallel converter, a pair of bit-bybit gate, a pair of 4-bit latch, and a pair of optional format
adapter sections.
Figure 3. Principle of QAM receiver
QAM mapper and demapper designs implemented in this
research refer to those models. For adaptive modulation, the
mapper and demapper must be configurable. In this research,
they can be configured to form as a 4, 16, 64, or 256 orders.
Figure 2 and 3 are generic form. For real applications,
additional sections may be inserted to improve the system
performance.
ISSN: 2085-6350
Figure 6. Block diagram of the implemented configurable QAM mapper
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
63
A. Signal Flow Part
Following the S/P, there are two identical arms: in-phase
(I) and quadrature (Q). The main signal flow part will be
elaborated further using Figure 7.
Figure 9. I-arm 2 MSBs gate section output for 256-QAM
Figure 7. Signal flow part, from S/P to latch section
The bit-by-bit S/P employs an 8-bit serial-in parallel-out
(SIPO) shift register 74LS164 preceded by an 74LS74 D-FF
for synchronizing the input bits. According to Figure 7, the
nth-stage output bit is the (8 – n) bits delayed version of the
incoming serial bit. The first stage S/P output bit (b7) is
shown in Figure 8. The strobe pulse (Stb) displayed in this
figure is for 64-QAM. This figure also shows that bits are
shifted at the beginning of their duration.
Figure 10. I-arm 2 LSBs gate section output for 256-QAM
Figure 8. Bit clock, 64-QAM strobe pulse, input bit stream, and the first
stage S/P output bit
Bit-by-bit gate section consists of four AND gates using
IC 74LS08. Each bit coming from S/P is AND-ed with its
associate gating logic (G) generated by the control circuit,
according to the selected QAM order. Operation performed
by this section is illustrated in Table I. As examples, Figure
9 to 12 describe Table I for 256-QAM. The highest delay is
eight bits long, which is experienced by the MSB (b0), due to
the S/P operation as explained before.
Figure 11. Q-arm 2 MSBs gate section output for 256-QAM
TABLE I. OPERATION OF THE BIT-BY-BIT GATE SECTION
QAM
Order
4
16
64
256
G3
1
1
1
1
Gating Logic
G2
G1
G0
0
0
0
1
0
0
1
1
0
1
1
1
B3
b0
b0
b0
b0
Output Bits
B2
B1
0
0
b2
0
b2
b4
b2
b4
B0
0
0
0
b6
Figure 12. Q-arm 2 LSBs gate section output for 256-QAM
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
64
Proceedings of CITEE, August 4, 2009
The two 4-bit latch (IC 74LS374) sections provide valid
bit (parallel bits) immediately after strobe. As listed in Table
I, the un-gated bit (bits) will also be presented as 0 at the bitby-bit gate sections output, and so at the latch output. As
examples, Figure 13 and 14 show the latched desired bits
(D3 – D2) output for 16-QAM.
Figure 9 to 14 also show that strobe pulse arrives at the
second half of the bit duration. Therefore, the latch sections
generate an additional delay of half bit duration.
The 4-bit parallel data format from these latches may not
be match with the format required by the IFFT section.
Therefore, a format adapter is necessary. Such required
formats may be unsigned number, signed number, 2’s
complement, or others. In this research, the format adapter is
a proportional expander converting the latch output to 8-bit
format. Its operation principle is shown in Figure 15 and
Table II.
For 4, 16, and 256 orders, these proportional expansions
are perfectly performed. For 64-order, the 8-bit format is
obtained by puncturing its 9-bit LSB. However, a slight
distortion occurs. The seven multiplexers in Figure 15 are
implemented using four dual multiplexers IC 74LS153.
Reason of employing magnitude expander is explained
as follows. In fact, the latch section output word-length is 1,
2, 3, or 4 for 4, 16, 64, or 256 orders, respectively. In other
words, its magnitude range is depended on the selected
QAM order, due to the difference in the number of bits. The
higher the order, the larger the range is. If the latch output
like this is fed directly to the IFFT block, then this block
must be adjusted according to the selected QAM order.
Briefly, it needs an ‘adjustable’ IFFT block. This scheme is
impractical, because increasing this block complexity.
Figure 13. I-arm latched desired bits for 16-QAM
By expanding the N-bit word-length to its proportional 8bit one, the dynamic range seen by the IFFT block is equal,
regardless the selected QAM order.
B. Control Circuit Part
The control circuit part will be elaborated further using
Figure 16. The selected QAM orders are determined by the
Y1Y0 bits. The gating logic encoder operation is described
by Table III, Figure 17, and Figure 18.
Figure 14. Q-arm latched desired bits for 16-QAM
Figure 16. Block diagram of the control circuit part
TABLE III. OPERATION OF THE GATING LOGIC ENCODER
QAM
Order
4
16
64
256
Figure 15. Principle of the 4-to-8 bit format adapter
TABLE II.
QAM
Order
4
16
64
256
q7
B3
B3
B3
B3
ISSN: 2085-6350
OPERATION OF THE FORMAT ADAPTER
q6
B3
B2
B2
B2
Format Adapter Output
q5
q4
q3
q2
B3
B3
B3
B3
B3
B2
B3
B2
B1
B3
B2
B1
B1
B0
B3
B2
q1
B3
B3
B3
B1
q0
B3
B2
B2
B0
Select
Y1
0
0
1
1
Y0
0
1
0
1
G3
1
1
1
1
Gating logic
G2
G1
0
0
1
0
1
1
1
1
G0
0
0
0
1
According to this table, two NOT gates (IC 74LS04),
four AND gates (IC 74LS08), and three OR gates (IC
74LS32) are needed to realize it.
Frequency divider is implemented using a cascaded two
4-bit binary counters (IC 74LS161) and produces two clock
signals with frequency as high as 2fb and fb, respectively.
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
Synchronization, if necessary, is possible by clearing the
counters momentary at the synchronizing pulses timing
instants. Figure 19 shows an example of the exciting and
produced clock signals. Figure 9 to 14 informs that strobe
pulse is nothing other than the QAM symbol clock. It is
generated by a finite-state machine (FSM) producing the
required pattern shown in Figure 20.
65
strobes are decoded from S7 and S15, and the 256-QAM is
decoded from S15 only. Generating the 64-QAM strobe by
decoding the S11, of course, needs a special trick, different
from the others. Block diagram of this FSM is illustrated in
Figure 21. The 4-bit binary counter is IC 74LS161, and the
multiplexer is IC 74LS153.
Figure 20. Required strobe pulses timing diagram
Figure 17. Gating logic encoder operation (2 MSBs)
Figure 21. Block diagram of the strobe pulse generator
Timing relationships among the strobe pulse and other
signals have been shown in several previous figures. A
comparison of the four strobe signals is shown in Figure 22.
Figure 18. Gating logic encoder operation (2 LSBs)
Figure 22. Comparison of the four strobe signals
V. IMPLEMENTATION OF DEMAPPING TECHNIQUE
Figure 19. Example of exciting and produced clock signals
Figure 20 shows that the bit clock and QAM symbol
clock (strobe pulse) are appear at the bit beginning and
ending instants, respectively. These strobe pulses are formed
by decoding the FSM output for the appropriate states. The
strobe pulses for 4, 16, 64, and 256 orders arrive once at
every 4, 8, 12, and 16 states, respectively. Therefore, the
FSM must be able to produce 16 states: S0, S1, …, S15. For 4,
16, and 256 orders, this FSM is let to run freely. The 4-QAM
strobes are decoded from S3, S7, S11, and S15. The 16-QAM
The demapper must be compatible to the mapper. Block
diagram of the implemented configurable (4/16/64/256)
QAM demapper is shown in Figure 23. This diagram
indicates two differences with Figure 6. First, the parallel-toserial (P/S) converter not only needs bit clock, but also load
pulse. Second, the demapper operation timing must be
synchronized to the arriving in-phase (I’) and quadrature
(Q’) signals.
A. Signal Flow Part
The signal flow part block diagram is shown in Figure
24. If the format adapter at the mapper side is just an
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
66
Proceedings of CITEE, August 4, 2009
expander, as implemented in this research, then the format
adapter in the demapper must be a compressor. It does not
need any electronic circuit, because it just takes the four
MSBs and leaves the remainder four LSBs.
Figure 23. Block diagram of the implemented configurable QAM
demapper
The P/S is implemented using an 8-bit parallel-input
serial-output (PISO) shift register 74LS165. Relation
between load pulse and serial output at P/S (for 256-QAM)
is shown in Figure 27.
Figure 26. Bit-by-bit gate parallel output bits for 16-QAM when its
parallel input is 1111
Figure 27. Relation between load pulse and serial output (256-QAM)
Figure 24. Signal flow part, from latch to P/S sections
The 4-bit latch and bit-by-bit gate sections are exactly
similar to the other ones at the mapper, so those sections
need not be explained again. Figure 25 shows that the Ds
input bits are latched to Bs output bits at the strobe pulse
time instants.
B. Control Circuit Part
In addition to generating signals similar to the signals
generated by the mapper control circuit, the demapper
control circuit must also generate the load pulse and must be
synchronized to the arriving symbol. It is the half bit
duration shifted and inverted version of the strobe pulse, and
can simply be generated by a manner illustrated in Figure 28.
Figure 29 shows the relation between strobe and load pulse
for 256-QAM.
Figure 28. Load pulse generation
If the selected QAM order is changed dynamically, the
strobe pulse is also able to track this variation, as shown in
Figure 30. At a fix bit rate, the higher the QAM order, the
lower the strobe rate is.
Figure 25. Latching instants of the parallel input bits
The bit-by-bit gate section operation is illustrated in
Figure 26. In this figure, the selected QAM order is 16, and
all parallel input bits are set to 1. The two MSBs are gated,
and the remainder two LSBs are un-gated (set to 0 at output).
ISSN: 2085-6350
Synchronizing pulse is used to synchronize the demapper
bit clock, strobe, and load by resetting all of the running
counters. Therefore, the arriving instants of this
synchronizing pulse must be at the symbol transition
instants, and not necessarily at every transition time. The
synchronized bit clock and strobe pulse are shown in Figure
31 and 32.
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
67
VI. MAPPING-DEMAPPING PERFORMANCE
Mapper-demapper compatibility is observed by
arranging the equipments as shown in Figure 33. External bit
source generates either a regular or pseudorandom binary
stream (PRBS). The synchronization pulse generator
simulator is a real synchronization circuit which has also
been constructed in chain with research, but the discussion is
beyond the scope of this paper. Therefore, in Figure 33, this
circuit is represented by a dash-lined box.
Figure 29. Relation between strobe and load pulse for 256-QAM
Figure 33. Compatibility test set-up
Figure 30. Effect of QAM order dynamic variation to the strobe pulse
Figure 34. Demapper output bit stream compared with mapper input bit
stream for 256-QAM
Figure 31. Pre and post-synchronized bit clock
Figure 35. Demapper output bit stream compared with mapper input bit
stream for 64-QAM
Figure 34 shows the demapper output (recovered) bit
stream (bOUT) compared with the mapper input (original) bit
stream for 256-QAM.
Figure 32. Pre and post-synchronized strobe pulse
As explained above, the mapper parallel output is
delayed from its original input bit stream as long as 8.5 bits,
regardless what the QAM order is. In fact, this parallel-bit is
ready to be read by the demapper, because at the same time,
it has been available at the demapper input latch. But, to read
these available bits, demapper must wait for the next strobe
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
68
Proceedings of CITEE, August 4, 2009
pulse. In synchronized condition, for 256-QAM, this waited
pulse will appear again eight bits duration later. So, these
parallel bits are latched-in to the demapper 8.5 + 8 = 16.5
bits later than the mapper original serial bit input. At this
time, these bits are available at the demapper P/S input and
waiting for be loaded to the P/S. The waited load pulse is
generated half bits later. Therefore, the complete delay is
16.5 + 0.5 = 17 bits, as shown in Figure 34.
Another similar result, for 64-QAM, is shown in Figure
35. In 64-QAM, the waited next strobe will appear again six
bit durations later, so the complete delay becomes 8.5 + 6 +
0.5 = 15 bits.
It must be noticed that the demapper half bit duration
delay is just a pseudo-delay, provided that this mapperdemapper pair section is operated in a complete
configuration shown in Figure 4 and Figure 5. In such
complete configuration, this half bit delay According to
those explanations, it can be concluded that the demapper
does not generate latency. On the other hand, the mapper
really generates latency as long as 8.5 bits.
All observations described above are performed at about
80 kbps bit rate. The control circuits have also been tested to
operate at about 4 Mbps bit rate, and the result is shown in
Figure 36.
VII. CONCLUSION
A hardware model implementing a technique of QAM
mapping-demapping for an adaptive modulation OFDM
system has been constructed in this research. These models
are constructed as a learning-equipment. The operating
QAM order can be selected as 4, 16, 64, or 256.
The mapper maps the incoming serial bit to the
corresponding I-Q pair QAM symbol. Both I and Q symbols
are expanded to 8-bit parallel format. The demapper demaps
the recovered I-Q pair 8-bit parallel format back to serial bit
output. Due to the configurability property required, the
mapper and demapper control circuits may be best
implemented as an FSM. The mapper generates latency as
long as 8.5 bits duration. The 0.5 bit latency generates by the
mapper just a pseudo-latency, because it will not appear in a
complete OFDM configuration operation. This mapperdemapper model is also capable supporting the bit rate up to
about 4 Mbps.
ACKNOWLEDGMENT
Idea of implementing this hardware model is also
supported
by
Mr.
Mulyana,
a
member
of
Telecommunication and High Frequency Laboratory at the
Electrical Engineering Department, Faculty of Engineering,
Gadjah Mada University. Construction and testing of this
model is completed by Amirul Azis Ashidiqy, Ibnu Gholib,
and Nurcahyo Wibowo, students at this department.
REFERENCES
[1]
[2]
[3]
[4]
[5]
Figure 36. Bit clock and strobe pulses for 4 Mbps bit rate
Figure 36 also shows that the signal waveforms are
seriously distorted, even though the levels are still in the
TTL permitted range. Nonetheless, operation in this rate is
possible.
ISSN: 2085-6350
[6]
[7]
Wilson, S.G., 1996, Digital Modulation and Coding, Prentice-Hall
International, Inc., Upper Saddle River, New Jersey.
Leon-Garcia, A., and I. Widjaja, 2004, Communication Networks
Fundamental Concepts and Key Architectures, McGrawHill, New
York.
Orthman, F., 2005, WiMAX Handbook Building 802.16 Wireless
Network, McGraw-Hill Companies, Inc., New York.
Schulze, H., and C. Luders, 2005, Theory and Applications of OFDM
and CDMA Wideband Wireless Communications, John Wiley &
Sons, Ltd., Chichester.
Stuber, G.L., J.R. Barry, S.W. Mclaughlin, Y.G. Li, M.A. Ingram,
and T.G. Pratt, 2004, Broadband MIMO-OFDM Wireless
Communications, Proceedings of the IEEE, Volume 92, Number 2,
February 2004.
van Nee, R., and R. Prasad, 2000, OFDM for Wireless Multimedia
Communications, Artech House Publisher, London.
Mulyana and B. Setiyanto, 2008, Implementasi Model PerangkatKeras Pemeta-Pengawapeta Bit-QAM dan Modulator-Demodulator
I/Q OFDM, Laporan Penelitian DIKS, Fakultas Teknik UGM,
Yogyakarta.
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
69
Hardware Model Implementation of a
Baseband Conversion, Chip Synchronization,
and Carrier Synchronization Technique
for a Universal QAM System
Budi Setiyanto, Mulyana, and Risanuri Hidayat
Electrical Engineering Department, Faculty of Engineering, Gadjah Mada University
Jl. Grafika 2 Yogyakarta, 55281, Indonesia
[email protected]
Abstract—Baseband converters, chip synchronizer, and
carrier synchronizer sections hardware model of a universal
quadrature amplitude modulation (QAM) system have been
implemented in this research. They are developed as a part of
an adaptive modulation orthogonal frequency division
multiplexing (OFDM) learning-equipment. Therefore, all
electronic components are selected from the general purpose
types, and performance quality is beyond the target. The
transmitter baseband converter is a digital-to-analog (DAC)
circuit. The receiver baseband converter is a cascaded
integrate-and-dump (I/D), sample-and-hold (S/H), and analogto-digital converter (ADC) circuit controlled by a triggered
non-cyclic finite state machine (FSM). Chip synchronization is
carried-out by detecting the abrupt amplitude changes of the
received signal. Synchronizing the carrier phase is carried-out
by shortly speeding-up or slowing-down the receiver local
voltage controlled oscillator (VCO), provided that no frequency
offset occurs. The ADC component limits the chip rate as high
as about 10 kcps.
Keywords—quadrature modulation, DAC, ADC, chip
synchronization, carrier synchronization
I. INTRODUCTION
Due to its compromise bandwidth efficiency and noiseimmunity [1], QAM is (and will be) widely applied in recent
(and future) wired and wireless technologies, either in single
or multi carrier system [2 – 6]. Examples of application
employing QAM are terrestrial microwave transmission and
almost all of orthogonal frequency division multiplexing
(OFDM)-based systems.
The QAM system baseband converter, chip synchronizer,
and carrier phase synchronizer sections have been
implemented in this research. Further, these constructed
sections are prepared to be integrated with the other sections
to form a complete adaptive modulation OFDM model. The
model will be packed as a learning-equipment. Therefore, all
electronic components are selected from the general purpose
types, and performance quality is beyond the target.
signals, modulate two orthogonal carrier signals, and finally,
transmitted as a QAM signal, s(t). Receiver-side recovers the
bit stream from the received QAM signal, r(t).
Figure 1. General form of QAM transmitter and receiver sides
Various processes transforming bits to chip and vice
versa are beyond the scope of this research. Therefore,
Figure 1 represents them as a dash-lined box. For single
carrier QAM, this process just organizes the serial bits to
their corresponding chip. For OFDM system, this process is
complicated.
The receiver attempts to extract or estimate I(t) and Q(t)
from the received signal, r(t), by carrying-out the
correlations
1 kT
I ' (t ) = Iˆ(t − kT ) =
r (t ) cos(2πf t )dt
∫
0
T (k − 1)T
and
Q' (t ) = Qˆ (t − kT ) =
II. REVIEW OF THEORETHICAL BACKGROUND
General form of a QAM transmitter and receiver sides
can be modelled in Figure 1. At the transmitter-side, the
input bit stream, b(nTb), is transformed to chip symbols
M(kT) = Re(kT) - jIm(kT), with Tb and T = NTb represent the
bit and chip durations, respectively. In general, each chip of
Re and Im symbols consists of N parallel bits. These chip
streams are converted to I(t) and Q(t) analog baseband
(1.a)
1 kT
r (t ) sin (2πf t )dt
∫
0
T (k − 1)T
(1.b)
Optimum estimation result is achieved when during the
integration interval, those two carrier signals are orthogonal
each other. This orthogonal property is guaranteed when f0 =
k/T (k = 1, 2, 3, …).
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
70
Proceedings of CITEE, August 4, 2009
III. IMPLEMENTATION OF BASEBAND CONVERSION
TECHNIQUE AT THE TRANSMITTER-SIDE
In Figure 1, baseband converter at the transmitter-side is
represented simply as a digital-to-analog converter (DAC)
only. There are two identical converters, for I (in-phase) and
Q (quadrature) arms. The complete configuration of each
converter is shown in Figure 2. For generality, the Re and Im
symbols are represented as D(kT), and the I and Q signals
are represented as a(t). The 8-bit DAC (IC 0800) is followed
by a current-to-voltage (I/V) converter. Level adjuster (LA)
is provided to adjust the final analog output signal range and
level. I/V and LA are based on operational amplifiers LF356.
Figure 5. Both converters output signal when the input is a free running
4-bit binary counter output
Figure 2. Configuration of baseband converter at the transmitter-side
Figure 3 shows an example of the DAC output voltage
(v(t)) when its first three MSBs are varied and the others are
fixed to be in zero state. This figure informs that the DAC is
able to track the input variation.
IV. IMPLEMENTATION OF THE BASEBAND CONVERSION
TECHNIQUE AT THE RECEIVER-SIDE
At the receiver-side, baseband converter also consists of
signal flow part and control circuit as shown in Figure 6.
Figure 6. Block diagram of the baseband converter and multiplier
Figure 3. Effect of input bits variation to analog output voltage of the
DAC-I/V section
Effect of level adjustment on the final converter output
voltage is shown in Figure 4. The LA can be exploited as a
level shifter, an amplitude scaling (attenuation or
amplification), or both.
A. Signal Flow Part
The correlator and ADC in Figure 1 can be explored
further as shown in Figure 7. Correlator section consists of
multiplier and integrate-and-dump (I/D) circuits. The
multiplier section is beyond the scope of this research, and
has ever been implemented [7]. Baseband converter consists
of I/D, sample-and-hold (S/H), level adjuster (LA), and
ADC circuits. There are also two identical converters, for I
and Q arms.
Figure 7. Correlator and ADC sections
The I/D circuit is desired to produce
Figure 4. Effect of level adjustments on the converter output
Both converters linearity and dynamic response have
been observed by driving the DAC with the output of a 4-bit
binary counter. The result is shown in Figure 5.
ISSN: 2085-6350
t
b(t ) = A ∫ u (τ )dτ , kδ < t < k (δ + T )
kδ
(2)
where A is a constant, k is an integer, and T is the chip
duration, so kδ is the instants when the dump pulse arrives.
The S/H samples b(t) at t = kσ, where kσ is the instants
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
71
when sample pulse arrives, and holds it until the next
consecutive sample pulse. Thus,
c(t ) = b(kσ )
(3)
Referring to equation (2), optimal result is obtained when
δ =0, and so does σ. Realizing δ =0 and σ = 0 is possible,
but impractical. In this research, δ =0 and σ = 0-, implying
As shown in Figure 9 and 10, sample and dump pulses
are bipolar (positive and negative) voltage.
Level adjuster (LA) is an inverting differential amplifier
circuit using operational amplifier LF 356. Operation
performed by this section is demonstrated by the signal
waveforms as shown in Figure 11.
that the integration result is sampled momentary prior to the
next consecutive integration period.
Principles of I/D and S/H circuits are illustrated in Figure
8, left and right part, respectively. Operational amplifiers
used in this figure are LF355. Operation performed by the
I/D section is demonstrated by the signal waveforms as
shown in Figure 9. This figure informs that integration is
started at the dump pulse instant.
Figure 11. Signal waveforms related to level adjuster (LA) operation
Figure 8. Integrate-and-dump and sample-and-hold circuits
This research employs IC 0808 successive approximation
register (SAR) type ADC. This IC needs a start pulse to
trigger convertion and clock. ADC starts converting at the
start pulse instant. The ADC end-of-conversion (EOC) status
signal goes high when once conversion has finished.
Operation performed by this section is demonstrated by the
signal waveforms as shown in Figure 12 and 13.
Figure 9. Signal waveforms related to I/D operation
Operation performed by the S/H section is demonstrated
by the signal waveforms as shown in Figure 10. This figure
informs that integrator output is sampled momentarily prior
to the next integration interval.
Figure 12. ADC clock, start pulse, and EOC status
Figure 13. Signal waveforms related to ADC operation
Figure 10. Signal waveforms related to S/H operation
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
72
Proceedings of CITEE, August 4, 2009
The ADC is followed by an 8-bit buffer 74LS373. The
buffer holds the last ADC output until refreshed by the next
consecutive result, as shown in Figure 14.
Figure 14. Signal waveforms related to buffer operation
Ideally, end-to-end baseband conversion test should be
carried-out by building connecting the DAC output to the
ADC input, exciting the DAC with a fast-varying 8-bit
parallel word, and observing the 8-bit ADC output. This
procedure required a 16-traces oscilloscope as a measuring
instrument. This instrument has not been available.
Figure 16 and 17 inform the ADC-DAC couple operate
properly, indicated by the DAC output which is able to track
the ADC input. The slower the ADC input varies, the more
similar the DAC output is, of course. In these tests, the ADC
is forced to operate at its maximum conversion rate as high
as 10000 conversions per second as stated in data-sheet.
B. Control Circuit
Block diagram of control circuit is shown in Figure 18.
Crystal oscillator generates 4.9 MHz clock. Frequency
divider is a free running 4-bit binary counter. It divides the
4.9 MHz exciting clock to produce 1.225 MHz ADC clock
and 306.25 triggered counter clock. Triggered counter is an
up-and-stop 4-bit binary counter. It starts when triggered and
stops when reaches its maximum state. The incoming
external triggering pulse may be too wide for directly
triggering the counter. Pulse shaper transforms this pulse to
an as narrow as desired pulse. Sample, start, and dump
pulses are formed by decoding three consecutive states, S0,
S1, and S2, respectively.
Without loss of significance, a simpler procedure is
created by reversing the configuration, as shown in Figure
15. The results are shown in Figure 16 and 17.
Figure 15. ADC-DAC compatibility test set-up
Figure 18. Block diagram of the control circuit
The frequency divider and triggered counter are
implemented using IC 74LS161. Figure 19 shows the clock
signals provided by frequency divider.
Figure 16. ADC-DAC link test result for 100 Hz sinusoidal signal
Figure 19. Frequency divider available outputs
External supplied triggering pulse may be too wide to be
directly fed to the triggered counter. Therefore, a pulse
shaper is inserted to form this pulse as narrow as desired.
The pulse shaper is a cascaded two monostable
multivibrators (MMV) IC 74LS123. Pulse position and
width are adjusted by the first and second stages MMV,
respectively.
Figure 17. ADC-DAC link test result for 1 kHz signal
ISSN: 2085-6350
External trigger, sample, start, and dump pulses are
shown in Figure 20. Sample and dump pulses shown in this
figure are in unipolar (zero and positive polarities) format.
As shown in Figure 9 and 10, these two pulses should be in
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
bipolar (negative and positive polarities) format. Therefore, a
unipolar-to-bipolar converter (UBC) circuit is inserted. UBC
circuit is simply an analog comparator circuit.
73
Relation between the received OFDM-signal with the
synchronizing pulse is shown in Figure 23. This pulse is
further used for synchronizing, not for exciting, the other
related circuits. Therefore, the appearance instants of this
pulse must be the chip transition instants, and not necessary
be at every chip transition. On the other, inter chip-transition
appearance is prohibited.
Figure 20. Trigger, sample, start, and sample pulses
V. IMPLEMENTATION OF CHIP SYNCHRONIZATION
TECHNIQUE
Example of a QAM signal waveform is shown in Figure
21 [7]. This figure shows that at every chip transition time,
signal amplitude and/or phase changes are very possible.
Figure 23. Chip synchronized pulse compared to the simulated received
OFDM signal
Above explanation informs that the chip-transition
detection like this is also applicable for general non constantamplitude modulation systems.
VI.
IMPLEMENTATION OF CARRIER SYNCHRONIZATION
TECHNIQUE
Referring Figure 1, adopting equation (1.a) and (1.b) for
estimating the real and imaginary parts chip gives
Figure 21. Examples of a QAM signal waveform
In this research, chip synchronization is performed by
searching for the abrupt change of received signal amplitude.
The pulse synchronizer generates a narrow pulse when such
change is detected. Block diagram of this chip synchronizer
is shown in Figure 22.
Figure 22. Block diagram of the chip synchronizer circuit
The envelope detector is a diode-capacitor circuit, similar
to one which is used in AM radio broadcasting receiver. The
differentiator is a capacitor-resistor (C-R) circuit, and its
output contains the dominant desired pulse contaminated by
other ripple pulses. Level threshold must be adjusted to
discard these ripples and extract the desired pulse. Level
threshold adjuster is just an analog comparator circuit. The
pulse width is adjusted by a pulse shaping circuit. The pulse
shaper is implemented using a monostable multivibrator IC
74LS123.
R̂e(t − kT ) =
1 kT
r (t ) cos(2πf t )dt
∫
0
T (k − 1)T
Îm(t − kT ) =
1 kT
r (t ) sin ( 2πf t )dt
∫
0
T (k − 1)T
(3.a)
and
(3.b)
Equation (3.a) and (3.b) inform that estimation requires a
synchronous local oscillator signals: cos (2πf0t) and sin
(2πf0t).
Carrier recovery technique [8] commonly used in M-ary
modulation system principally is adoptable for QAM if the
chip word-length (number of bits per chip) is known by the
receiver. For OFDM system, this requirement is unrealistic,
because it means that the receiver must know the number of
bits per point of the transmitter IFFT output. Therefore, this
research proposes a carrier synchronization technique which
is independent to the chip word-length. This technique will
be explained as follows.
Let the receiver has been successful to generate the
desired frequency. The next problem is synchronizing its
phase. In usual practical application, a special time-slot is
provided where a specific real and imaginary parts pattern
are transmitted to guide the receiver in synchronizing its
phase. Using signal identification, the receiver know whether
its synchronous condition has been reached or not yet.
The proposed technique in this research is illustrated in
Figure 24. The voltage controlled oscillator (VCO) is
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
74
Proceedings of CITEE, August 4, 2009
controlled by a positive, zero, or negative voltage through an
analog multi-plexer. The control logic is generated by
another smart circuit responsible for signal identification as
explained above. This smart circuit is not presented here, and
beyond the scope of this research.
Figure 24. Principle of the carrier phase synchronization
Control circuit must know whether the VCO is being
lead or lag. Delaying or advancing the VCO is performed by
shortly slowing it down or speeding it up, that is by driving it
with a negative or positive voltage, respectively. The VCO is
left to be in free running condition by driving it with zero
control voltage.
Preliminary experiment using IC 4046 has shown that
varying the control voltage from -180 mV to +180 mV
varies almost linearly the frequency from 450 kHz to 550,
with a free running or centre frequency of about 500 kHz.
The VCO output waveforms are shown in Figure 25.
cascaded integrate-and-dump, sample-and-hold, analog
signal level adjuster, and 8-bit analog-to-digital converter
(ADC) with its suitable control circuits. Chip
synchronization is carried-out by detecting the abrupt
changes in the received signal amplitude. This section
consists of envelope detector, differentiator, threshold level
adjuster, and pulse shaper circuits. Carrier synchronization is
performed by adopting the delay-locked loop (DLL)
technique commonly employed in a spread-spectrum system
pseudorandom code tracking circuit. This section consists of
a voltage difference detector, a lowpass filter (LPF), and a
voltage controlled oscillator (VCO). These sections are
implemented as a trainer, and operate at the chip rate as high
as about 10 kcps. The chip rate and carrier frequency are
constrained mainly by the ADC and VCO, respectively.
ACKNOWLEDGMENT
Idea of implementing this hardware model is also
supported by Mr. Astria Nur Irfansyah, a member of Basic
Electronics Laboratory at the Electrical Engineering
Department, Faculty of Engineering, Gadjah Mada
University. Construction and testing of this model is
completed by Amirul Azis Ashidiqy, Ibnu Gholib, and
Nurcahyo Wibowo, students at this department.
REFERENCES
[1]
[2]
[3]
[4]
[5]
Figure 25. VCO output at various control voltage level
VII. CONCLUSION
[6]
A hardware model implementing a technique of
baseband conversion, chip synchronization, and carrier
synchronization for an OFDM system has been constructed
in this research. At the transmitter-side, the baseband
converter section is a pair of cascaded 8-bit digital-to-analog
converter (DAC) and analog signal level adjuster circuits. At
the receiver-side, this baseband converter is a pair of
ISSN: 2085-6350
[7]
[8]
Wilson, S.G., 1996, Digital Modulation and Coding, Prentice-Hall
International, Inc., Upper Saddle River, New Jersey.
Leon-Garcia, A., and I. Widjaja, 2004, Communication Networks
Fundamental Concepts and Key Architectures, McGrawHill, New
York.
Orthman, F., 2005, WiMAX Handbook Building 802.16 Wireless
Network, McGraw-Hill Companies, Inc., New York.
Schulze, H., and C. Luders, 2005, Theory and Applications of OFDM
and CDMA Wideband Wireless Communications, John Wiley &
Sons, Ltd., Chichester.
Stuber, G.L., J.R. Barry, S.W. Mclaughlin, Y.G. Li, M.A. Ingram,
and T.G. Pratt, 2004, Broadband MIMO-OFDM Wireless
Communications, Proceedings of the IEEE, Volume 92, Number 2,
February 2004.
van Nee, R., and R. Prasad, 2000, OFDM for Wireless Multimedia
Communications, Artech House Publisher, London.
Mulyana and B. Setiyanto, 2008, Implementasi Model PerangkatKeras Pemeta-Pengawapeta Bit-QAM dan Modulator-Demodulator
I/Q OFDM, Laporan Penelitian DIKS, Fakultas Teknik UGM,
Yogyakarta.
Ha, T. T., 1990, Digital Satellite Communications, 2nd ed., Mc-Graw
Hill, Singapore.
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
75
Comparison Study of Breast Thermography and
Breast Ultrasonography for Detection of Breast
Cancer
Thomas Sri Widodo1), Maesadji Tjokronegore2), D. Jekke Mamahit3)
1) &3)
Electrical Dept., Eng. Fac., Gadjah Mada University
2)
Medicine Fac., Gadjah Mada University
Abstracti-Breast thermography is a non-invasive diagnostic
technique that allows the examiner to visualize and quantify
changes in skin surface temperature. The objective of this
research is to compare breast thermography and breast
ultrasonography during abnormal physiology monitoring period
for early stage cancer detection.
For this research the breast thermogram of four patients
are compared with breast sonogram of the same patients with
known breast diseases. The results show that thermography can
be used to detect breast cancer with all stages with exception of
cysts disease in the breast.
Key words: thermography, sonopgraphy, breast, cancer
I. Introduction
Thermography may be defined as a method of mapping
surface temperature variations that depends entirely upon
radiated energy in the infrared spectrum. A detector within
the apparatus converts infrared radiation to an electrical
output proportional to the emitted radiation [1].
The procedure is based on the principle that the
chemical and blood vessel activity in both pre-cancerous
tissue and the area surrounding a developing breast cancer
is almost higher than in the normal breast. Since precancerous and cancerous masses are highly metabolic
tissues, they need an abundant supply of nutrients to
maintain their growth. In order to do this they increase
circulation to their cells by sending out chemicals to keep
existing blood vessels open, recruit dormant vessels and
create new ones (neoangiogenesis). This process results in
an increase in regional surface temperatures of the breast [2,
3].
Criteria for the abnormal breast thermogram may be
derived into vascular and nonvascular thermal patterns. The
number of vessels should be relatively equal bilaterally .
Vessels that are tortuous or serpiginous, or tend to cluster
should be considered abnormal. Nonvascular criteria
include the edge sign, which is an asymmetric flattened
breast contour or localized protrusion with in the breast.
Since there is a high degree of thermal symmetry in the
normal body, subtle abnormal temperature asymmetry can
be easily identified [1, 2].
II. Category and Stage of Tumor
There are five thermo-biological (TH) categories for
breast thermogram. [4]:
TH1 - Normal uniform non vascular.
TH2 - Normal uniform vascular. There
symmetrical streaks of thermal findings (blood vessels) in
both breast. This increase in metabolism is caused by a
relative progresterone deficiency (estrogen dominance).
TH3 – Equivocal (questionable). There is single hot blood
vessel in one breast (right or left). This thermal finding will
need to be watched over time for change. If it is remain
stable, or improves, this finding is normal for the
physiology of the patient.
TH4 – Abnormal. There is an increased temperature area
(metabolism) of one breast. Especially significant is the
highly vascular area.
TH5 – Severely abnormal. There is an intensely increased
temperature (metabolism) of the entire one breast. This is
possible an inflammatory carcinoma.
The currently accepted staging scheme for breast cancer
is the TNM classification. The five tumor T classification
values are Tis, T1, T2, T3, and T4 [5].
Tx – Primary tumor cannot be
assessed.
T0 – No evidence of primary tumor.
Tis - Carcinoma in situ.
T1 - Tumor 2 cm or less in its
greatest dimension
T2 - Tumor more than 2 cm but no
more than 5 cm in its greatest
dimension.
T3 - Tumor more than 5 cm in its
greater dimension.
T4 - Tumor of any size with direct
extension to chest wall or skin.
III. Comparison Results with Breast Ultrasonography
Thermogram images of four patients with different
stages were imaged using an infrared camera (Fluke
Thermal Imager Model I 120) in an hospital. These images
are then compared with the ultrasonography images
(sonogram) of the same patients which are shown in Figure
1 to 4.
IV. Conclusion
There is correlation between thermo biological breast
category and breast cancer stage
for breast cancer
classification. Non uniform thermal findings in both breast
indicates a breast cancer in one breast. This comparison
study shows that breast thermography can be used to detect
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
76
Proceedings of CITEE, August 4, 2009
a breast cancer. Cyst in the breast can not be detected using
breast thermography.
Figure 3. (a) There is an increased temperature area (metabolism) of left
breast (TH4). (b). Left breast sonogram shows benign tumor without
malignancy (T2).
Figure 1. (a) Entirely breast thermogram shows normal uniform
nonvascular (TH1). (b) and (c) Right and left breast sonogram show small
cysts (T0).
Figure 2. (a) Entire breast thermogram shows a thermal finding at the
upper part of the right breast (TH3). (b) Right breast sonogram shows a
mass of disfuse dextra sinistra of 0.8x0.4 cm (T1) with malignancy and
metastasis.
Figure 4. (a) There is an intensely increased temperature (metabolism) of
the entire left breast (TH5). (b) Left breast sonogram shows a malignancy
tumor with micro calcification of 3.5 cm (T3).
ISSN: 2085-6350
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
References
[1]. Carl J.D’Orsi,MD & Richard E.Wilson,MD, 1983,
Carcinoma of the breast : Diagnosis and
Treatment, Boston/Toronto
[2]. Anonium, 2003., Breast Health,
Meditherm Inc
[3]. Gros C., Gautherie M., 1980, Breast Thermography
and Cancer Risk Prediction, Cancer
77
[4].
Anonymous,
2007,
Breast
Thermography,
[email protected].
[5].
Anonymous, 2009, Breast Cancer Classification,
Wilkipedia
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
78
Proceedings of CITEE, August 4, 2009
Adaptive Polynomial Approximation for
Gravimetric Geoid: A case study using EGM96 and
EIGEN-GL04C Geopotential Development
Tarsisius Aris Sunantyo
Geodetic Department, Faculty of Engineering, Gadjah Mada University.
55000, Indonesia
[email protected]
Muhamad Iradat Achmad
Electrical Department, Faculty of Engineering, Gadjah Mada University
55000, Indonesia
[email protected]
Abstract— In this paper, adaptive polynomial approximation
method to model the gravimetric geoid is presented. The
polynomial approximation model to arrange training of input
is for adaptive system. Training data pairs (input and output)
were compiled from latitude and longitude data sequences as
input training, and an associated geopotential developments
datum on each spatial position as output training. By preceded
centering of latitude and longitude data, input training formed
with following appropriate formula is the polynomial terms.
Adaptation process used LMS (least mean square) algorithm
in weight updating, and after training session, approximation
of desired target was computed by reloading weights into the
polynomial model. Model of assessment test used was to
validate adaptive model by comparing residual distance of
consecutive point from both geopotential developments data
and respective adaptive model in geocentric coordinates
system. Using geopotential developments data around of
Merapi and Merbabu as a case study, the results show that the
residual distance between geopotential developments data and
respective adaptive model are about 0.0014417 m (in total
absolute value with standard deviation is about 4.6906x10-5 m)
using EGM96 geoid and about 0.0014468 m (in total absolute
value with standard deviation is about 4.702x10-5 m) using
EIGEN-GL04C geoid. Better result could be achieved by
adding more training session with smaller gain factor.
Keywords—
Adaptive
polynomial
approximation,gravimetric geoid, geopotential developments, and model
assesment.
I. INTRODUCTION
One of the fundamental problems in geodesy is to define
the shape and size of the earth and its gravity field taking
into account its temporal variation. Representation of the
shape of the earth can be carried out by several methods;
one of them is by geoid. Geoid is an equipotential surface of
the earth at mean sea level (Heiskanen and Moritz, 1967).
The use of precise geoid related data, in particular its
undulation, is widespread in all branches of geodesy and it
is often analyzed in other Earth sciences for example in
geophysics, oceanography, as well as in civil engineering
(Zhicai and Yong-qi, 2002).
ISSN: 2085-6350
Figure 1. Distribution of the Indonesian active volcanoes as the Ring of
Fire (Red color is active volcano) (Sunantyo, 2008)
Indonesia, as an archipelago, located partly on the
Eurasian plate, which is subducted by the three major
plates: the Indo-Australian plate in the south and in the
west; the Pacific plate in the East and the Philippine Sea
plate in the north. These subduction zones around Indonesia
create a ring of volcanoes, which is called “the Ring of
Fire” (see Figure 1). A total of 129 active volcanoes exist,
and one of them is Merapi as a result of the subductions.
Purbawinata et al.,(1997) state that the subduction zone is
marked by a chain of active and dormant volcanoes which
spreads along Sumatra, Java, Bali, Lombok, Sulawesi to the
eastern part of Indonesian. North of Merapi volcano is
Merbabu volcano which has not had activities more than
200 years. These two volcanoes are located in Yogyakarta
city and Central Java (Purbawinata et al., (1997)). These
two volcanoes are very often to be used as a research area
too many scientists (i.e. geophysicist, geologist, geodesist
etc) to have a comprehensive understanding about them.
Merapi volcano has been labeled as one of 15 high risk
volcanoes by the International Decade of Natural Disaster
Reduction program (IDNDR) of UNESCO (Sunantyo,
2008). The integration a multidisciplinary approach
(geodesy, geology, geophysics, seismology etc) concerning
to understand the structure and volcanic activities of Merapi
in more detail is very important and urgent (Zschau et al.,
1998). In physical geodesy aspect, gravimetric geoid
determination in these two volcanoes using geopotential
developments is discussed in this research.
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
79
[ ]
Both the estimation of Cost function ξ k = E ε k ≅ ε k used
2
2
in computing instant gradient at the kth steps, ∇ε k , and gain
factor µ , are used in weight updating
2
w k +1 = w k − µ∇ (ε k2 ) .
(5)
Gradient operator ∇(•) defined as column vector
⎡ ∂ (• )
∇(• ) = ⎢
⎣⎢ ∂w1,k
∂ (• )
∂w2 ,k
∂ (• ) ⎤
L
⎥ (6)
∂w p ,k ⎦⎥
T
and the ith element of gradient vector is
∂ (ε k2 )
∂ε k
= 2ε k
∂wi ,k
∂wi ,k
Figure 2. Mechanical block of current adaptive system (Widrow and
Stearns, 1985)
II. ADAPTIVE SYSTEM DESIGN
Input training of adaptive system is formed by following
associated formula in the polynomial terms. For the 4th order
of 2D polynomial (Sunantyo, 2008):
h = c0,0 + c1,0 λ + c0,1ϕ + c 2,0 λ 2 + c1,1 λϕ + c0,2ϕ 2 +
(1)
c3,0 λ + c2,1 λ ϕ + c1,2 λϕ + c0,3ϕ +
3
2
2
3
c 4,0 λ + c3,1 λ ϕ + c2,2 λ ϕ + c1,3 λϕ + c0,4ϕ .
4
3
2
2
3
4
where λ, φ and h are longitude, latitude, and geopotential
developments respectively, we have 15 terms (including
constant term), and 15 parameters. For convenience, we use
a1, a2,…,a14 to represent the terms: λ, φ, …, φ4 and w0, w1,
w2, …, w14 for parameters: c0,0, c1,0, c0,1, …, c0,4. By preceded
centering of latitude (=φ) and longitude (=λ) data, input
training vector arranged by following appropriate formula in
each polynomial terms. For example, the fifth term of kth
input training given by
(
)(
a4 k = λk − λ ϕ k − ϕ
)
(2)
The kth output of training data, hk, is an associated
geopotential developments datum on spatial position (λk, φk).
Mechanical block of current adaptive system shown in Fig. 1
where k is time index, system output fk denotes an
approximation of desired output zk, and residu εk gives the
difference between zk and fk. Weight updating that has a
basic form (scalar input vector), gives new weight vector
wk+1, and guarantees that adaptation process always parallel
to input vector. Output system at time index k is given by
fk =
L
∑
wik a ik
(5)
i =0
and residue estimation
ε k = zk − fk .
(4)
(7)
By using (4) in (7) and noting that fk is independent of wk,
we obtain
∂ (ε k2 )
∂f k
(8)
= −2ε k
∂wi ,k
∂wi ,k
Using (3) in (8), we get
∂ (ε k2 )
= −2ε k ai ,k ,
∂wi ,k
and gradient vector become
∇ (ε k2 ) = −2ε k a k
[
where a k = a1,k
a 2 ,k
L a p ,k
(9)
]
T
(10)
is input vector at time
index k. Substituting (10) in (5) we get
w k +1 = w k + 2 µ ε k a k .
(11)
This is referred to the LMS algorithm shown in Fig. 1 as the
dash line block. For N training data presentations, we define
input matrix
A = [a 1 a 2 K a n K a N ] T (12)
where n is data index and using it to estimate gain factor as
0.01
µ≅
.
(13)
trace(E[A T A])
where E[.] is expectation operator.
III. MODEL ASSESMENT TEST
Model assessment test was used to validate model. This
test compares the distance between two consecutive points
given by both geopotential developments data and its
adaptive model in geocentric system (Sunantyo, 2008).
Both residual distance (in total absolute value) and
respective standard deviation are used to asses the modeling
capability of adaptive system. For two consecutive points in
geodetic system, (λi ,ϕ i , hi ) and (λi +1 ,ϕ i +1 , hi +1 ) , the distance
(in meter) is given by
d=
(x
− xi ) + ( yi +1 − yi ) + (z i +1 − z i ) (14)
2
i +1
2
2
where
⎛ π ⎞ ⎛ π ⎞
x = (N ′ + h ) cos⎜ ϕ
⎟ cos⎜ λ
⎟,
⎝ 180 ⎠ ⎝ 180 ⎠
⎛ π ⎞ ⎛ π ⎞
y = (N ′ + h ) cos⎜ ϕ
⎟ sin⎜ λ
⎟,
⎝ 180 ⎠ ⎝ 180 ⎠
z = (N ′(1 − e 2 ) + h )sin ϕ ,
K
N′ =
,
1 − e 2 sin 2 ϕ
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
80
Proceedings of CITEE, August 4, 2009
K = 6378137,
e = 8.1819190842622x10-2.
Difference between the distance given by gravimetric
data dh and one other given by adaptive model df (called
residual distance) which defined as
∆d = d z − d a ,
(15)
and has a total value
∑ ∆d =
N −1
∑ ∆d
j
,
(16)
j =1
with standard deviation given by
∑(
)
1
2
2 ⎞
⎛ 1 N −1
σ ∆d = ⎜⎜
∆d j − ∆d ⎟⎟ ,(17)
⎝ N − 1 j =1
⎠
represents the closeness of model from its desired target.
Denominator (N-1) in (17) referred to the number of
elements of residual vector related to the consecutive points
selecting.
IV. IMPLEMENTATION
The two kinds of geopotential developments data are
EGM96 and EIGEN-GL04C, located in 48 different stations
around Merapi and Merbabu volcanoes. The datasets consist
of longitude, latitude, and an associated geoid height from
both EGM96 and EIGEN-GL04C. At the design stage, we
use (2) to center both longitude (= λ ) and latitude (= ϕ )
data, and arrange them following the polynomial terms in (1)
to be input training data, while in the output side we use two
sets of geopotential developments data to be our desired
target. Using (12), it was defined input matrix
A ∈ ℜ48×15 where 48 rows and 15 columns referred to the
number of data presentations and the number of weights
respectively. Substituting input matrix A in (13), then is was
got gain factor µ = 0.0095. Now, it was ready to develop an
algorithm to handle our modeling task.
Algorithm of gravimetric geoid modeling using adaptive
polynomial approximation is explained as follows:
Input:
Latitude (= ϕ ), longitude (= λ ), geopotential
developments height (=h), gain factor (= µ ), and
maximum iteration (=kmax).
end
4. Computing output vector f:
for n=1,2,…,48,
fn=anTwkmax
end
V. RESULT AND DISCUSSION
In the first part, distribution of TTG and its corresponding
geoid using EGM96 and EIGEN-GL04C were discussed,
while the second part analyzes are contour lines of geoid
model and difference between geoid data and its model of
EGM96 and EIGEN-GL04C. The third part is model of
assessment using EGM96 and EIGEN-GL04C model for
the 1D plotting of, data and model in the distance axis,
distance of consecutive points and residual distance.
Figure 2. Distribution of TTG and its corresponding geoid using EGM96.
The Figure 2. show that distribution of TTG TTG and
its corresponding geoid using EGM96 that all of TTG from
Yogyakarta, Muntilan, Surakarta are mostly up (plus) and
TTG from Surakarta to Yogyakarta are mostly down
(minus).
Output: Approximation of h (=f).
1. Centering ϕ , λ ∈ ℜ 48 :
ϕ c = ϕ − ϕ and λ c = λ − λ .
2. Arranging input training based on polynomial terms:
a n ∈ ℜ15 , n = 1,2,..., 48. .
3. Training:
k=0,w0=0
while k<kmax
for n=1,2,…,48
fn=anTwk
εn=hn-fn
wk+1=wk+2µεnan
end
k=k+1
ISSN: 2085-6350
Figure 3. Distribution of TTG and its corresponding geoid using EIGENGL04C
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
The Figure 3. show that distribution of TTG TTG and its
corresponding geoid using EIGEN-GL04C that all of TTG
from Yogyakarta, Muntilan, Surakarta are mostly up (plus)
and TTG from Surakarta to Yogyakarta are mostly down
(minus).
81
model gives smaller both residual distance and its standard
deviation than another geoid model, EIGEN GL04C.
Concerning with the Figures 2 and 3 that the distribution of
TTG TTG and its corresponding geoid using EGM96 and
EIGEN-GL04C are very similar (there is no significant
difference between them).
Figure 6. Model assesment using EGM96 model for
residual distance
Figure 4. Contour lines of gravimetric geoid model and difference between
geoid data and its model of EGM96
Figure 7. Model assessment using EIGEN-GL04C model for residual
distance
VI.
CONCLUSION
The conclusion of this research are summarized as
follows:
1.
Figure 5. Contour lines of gravimetric geoid model and difference between
geoid data and its model of EIGEN-GL04C
2.
Figures 4 and 5 show that the pattern of contour line of
gravimetric geoid height is very similar. It is supposed that
there is no significant difference contour line using EGM96
and EIGEN-GL04C. The geoid height between them are
also continue
Figures 6 and 7 show the results of model assesment
procedure related to the geoid model contoured in Figure 4
and 5, respectively. In that procedure, distance of
consequtive points in cartesian system are calculated from
both data and model. Residual distance which are defined as
difference distance between data and model are then used to
asses the model. From the results, we know that EGM96
3.
Adaptive polynomial approximation for
gravimetric Geoid height using EGM96 and
EIGEN-GL04C Geopotential Development is very
similar. It is supposed that there is no significant
difference contour line using EGM96 and EIGENGL04C.
The gravimetric geoid height using EGM96 and
EIGEN-GL04C are continue.
The distribution of TTG TTG and its
corresponding gravimetric geoid using EGM96
and EIGEN-GL04C are very similar
REFERENCES
[1]
[2]
Heiskanen, W. and Moritz, H., 1967, Physical Geodesy, Freeman,
San Francisco. Hofmann-Wellenhof B. And Moritz, 2006, Physical
Geodesy, Second Edition, Springer, Wien NewYork.
Pubawinata, M.A., Radomopurbo A., Sinulingga I. K., Sumarti S.,
and Suharno, 1997, Merapi Volcano A Guide
Book, The
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
82
[3]
[4]
Proceedings of CITEE, August 4, 2009
Volcanological Survey of Indonesia, Directorate General of Geology
and Mineral Resources, Bandung, Indonesia
Sunantyo, Modeling of Local Geoid using Adaptive Scheme: A Case
Study around Merapi and Merbabu Volcanoes, Central Jawa,
Indonesia, Doctoral Dissertation, Faculty of Engineering, Gadjah
Mada University, 2008.
Widrow, B., and Stearns, S. D., 1985, Adaptive Signal Processing,
Prentice Hall Inc.
ISSN: 2085-6350
[5]
[6]
Zachau J., Sukhyar, R., Purbawinata, M.A., Lühr, and Westerhaus,
M., 1998, Project MERAPI – interdisciplinary research at a high-risk
volcano, In : Decade–Volcanoes Under Investigation, Deutsche
Geophysikalische Gesellschaft, Sonderband III/1998.
Zhicai L. and Yong-qi C, 2002, Precise determination of Hongkong
Geoid using heterogeneous data, FIG XXII International Congress,
Washington, D.C. U.S.A.
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
83
Cerebellar Model Associative Computer (CMAC)
for Gravimetric Geoid study based on EGM96 and
EIGEN-GL04C Geopotential Development
1)
Muhamad Iradat Achmad1), Tarsisius Aris Sunantyo2)
Electrical Department, Faculty of Engineering, Gadjah Mada University
[email protected]
2)
Geodetic Department, Faculty of Engneering, Gadjah Mada University.
[email protected]
3)
Adhi Susanto3)
Electrical Department, Faculty of Engineering, Gadjah Mada University
Abstract— In this paper, cerebellar model associative
computer (CMAC) algorithm is proposed to be used for geoid
study. This algorithm based on the mapping scheme of input
space that transforms input similarity in the input space into
associated levels in the associated cells space. Similar inputs
mapped to similar levels, while dissimilar inputs associated to
mutual independent cells. Associated levels given by
overlapping base functions selected represent a subset of
“feature” to be parameterized using adaptive linear combiner
(ALC) structure where a set of “feature” impersonate as input
vector in the back side, parameter or weight vector in the
middle side, and an output in the front side. ALC structure
adjusts weights value using LMS (least mean square) recursion
in the training (or analyzing) session, and accumulates
weighted input to get an output in the synthesizing session.
Desired output using in weights adjustment are any values
from the output space that correspond to respective values in
the input space. Model assessment test used to validate CMAC
model by comparing residual distance of consecutive points
from both EGM96 and EIGEN-GL04C geopotential global
development is data and respective CMAC model in geocentric
coordinates system. By utilization of EGM96 and EIGENGL04C geopotential global development data sets acquired
from selected points of geometric geoid in around of Merapi
and Merbabu volcanoes, it is simulated the proposed
algorithm to study geoid. The results show that difference
between data and model based on EGM96 geopotential global
development are about 0.00026296 m in total absolute value
with standard deviation about 8.3856 x 10(6) m, and based on
EIGEN-GL04C geopotential global development is about
0.00027139 m in total absolute value with standard deviation
about 8.5441 m. These results use 1x106 epoch in training
session with gain factor mu=1.0638 x 10(4) m better result
could be achieved by adding more training session with
smaller gain factor.
Keywords—cerebellar
model
associative
computer
(CMAC), adaptive linear combiner, least mean square,
EGM96 and EIGEN-GL04C geopotential global development.
I.
INTRODUCTION
Gravimetric geoid study is a fundamental part in geodesy
and other earth sciences in case to define the shape and size
of the earth and its gravity field including its temporal
variation. Many applications such as EWARS (early warning
system) for disaster, oil and gas discovery, land
management, and many more, utilizate geoid information to
perform specific and dedicated operation in its
implementation. This is the facts to be the reason of our
study.
One of many tasks in gravimetric study is the modeling
of gravimetric geoid that has the same meaning with the
surface fitting in which two predicting variables positioned
as input and one response variable as the function of two
input variables to be an output. For the temporal variation of
the surface, we need to add an input time variable, but this is
not the case presented here. Related to the surface fitting,
there are many research available in text books and
references. Most of them use the generalizing of least square
method where prior knowledge given in input design
suspected as the characteristics of the two-dimensional (2D)
plant handled.
In this paper, we propose a gravimetric geoid modeling
method that differs with the previous method in the way to
incorporate prior knowledge to the modeling system. The
method uses an algorithm that exploits the generalization
and learning capability of CMAC (cerebellar model
associative computer, cerebellar model arithmetic computer,
or cerebellar model articulation controller). CMAC is an
associative memory neural network based on the mapping
scheme of input space that transforms input similarity in the
input space into associated levels in the associated cells
space. Similar inputs mapped to similar levels, while
dissimilar inputs associated to mutual independent cells.
CMAC is interpreted as a look up table method [1], and has
been used in the areas of control [2]-[6], pattern recognition
[7], signal and image processing [8]-[10]. Other research
related to the structure and learning of CMAC investigated
in [11]-[16].
In the proposed method, CMAC used to learn a non
linear mapping from a spatial position into the corresponding
geoid value. A geoid position is quantized by a number of
quantization levels, and the CMACs input taken from
quantizers output, associated to a set of association cells by
the input mapping scheme. Training (or analyzing) session
uses the association cells and the corresponding geoid value
to be a set of respective input and output training. By using
the LMS recursion, a set of weights adjusts until a number of
epoch reached. Syntesizing session accumulates the
weighted inputs to approximate desired output.
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
84
Proceedings of CITEE, August 4, 2009
II.
CMAC
In this section, we give a brief description of CMAC
with emphasis on generalization and learning capability. In
the proposed method, CMAC is used to learn a nonlinear
mapping from a spatial position into the corresponding geoid
value. Thus, we limit our discussion to a two-input singleoutput CMAC, although a general CMAC is capable of
learning a multi-input multi-output nonlinear function.
Input Mapping Scheme
Let the two-dimensional (2D) input of a nonlinear
function be (x,y), where x and y are a set of L real numbers,
{x,y}∈RL, in the closed interval, [xmin, xmax] and [ymin, ymax],
respectively. The input x is quantized by 1Q diferent uniform
quantizers with quantization interval 1Q, 1ui (i=0,1,2,…,U)
where 1u0≤xmin and 1uU≥xmax. The pre-superscript “1” used in
1
Q and in all symbols in this paper denotes the input
dimension. The decision level indexs of 1ui and 1ui+1 are
given by
prior the design. Equation (3) gives a number of g address
from overall address
p=
∑p
j
.
(7)
j
On each associated address, the association levels is
defined by
A.
⎛
⎞
− (g 2 4 )
⎟⎟
aki , j = exp⎜⎜ i
i
⎝ ( k − (Z a − 0.5 g ))((Z a + 0.5 g ) − k ) ⎠
(8)
where Za is the center of the covered inputs region in terms
of ik (i=1,2).
Adaptive Linear Combiner (ALC)
The structure of ALC is shown in Fig. 1. There is an
input vector, a set of adjustable weights, a summing unit, and
a single output. An input vector, a, with elements a1, a2, …,
⎢ x −1 u 0 1 ⎥
ap, has a number of g non zero elements where g is integer in
1
( Q )⎥ + 1
k =⎢1
(1)
1
the open interval (0,p). These inputs are parameterized by a
⎣ uU − u 0
⎦
corresponding sub set of adjustable weights, wj∈w, j=1, 2,
…, g, and summed up by the linear combiner to get a single
1
1
where k is a set of L positive integer in interval [1, Q]. The
output, y. The combiner is called “linear” because for a fixed
corresponding dequantization levels of (1) are given by
setting of the weights its output is a linear combination of the
input components. However, when the weights are in the
1
process of being adjusted, the output of the combiner is no
2( k ) + 1 1
~
(2) longer a linear function of input. Given an input vector to
x =1 u 0 +
× ( uU − 1 u 0 ) .
1
2( Q )
ALC be in the column form as
x is called the quantization error
Difference between x and ~
of the dimension of x. Similarly, the input y is also quantized
in the same way as in x. By changing in (1) and in (2), the
pre-superscript from “1” to “2”, the quantization interval
from “u” to “v”, and the index from “M” to “N”, we get a set
of L positive integer , 2k, in interval [1,2Q] and the
corresponding dequantization levels, ~y . Difference between
y and ~
y is called the quantization error of the dimension of
y.
B.
[
a n = a 1n
a 2n
L a pn
]
T
(9)
Let the two dimensional (2D) input of CMAC be (1k,2k)
where 1k and 2k are positive integers in [1,1Q] and [1,2Q]
respectively. A set of g address associated by these inputs is
given by
fa j = p j + (1 f j ) + (1 c j )(2 f j );
where
p j = (1 c j )( 2 c j ) ,
⎡ Q − dj ⎤
cj = ⎢
⎥ +1 ,
g
⎢
⎥
i
i
⎡
⎤
k
d
−
j
i
fj = ⎢
⎥ , i=1,2.
⎢ g ⎥
i
Figure 1. Adaptive linear combiner (ALC).
j = 1, K , g (3)
(4)
i
i
(5)
(6)
The parameters of CMAC, idj and g, are displacement and
generalization parameters, respectively. They are defined
ISSN: 2085-6350
In this notation T stands for transpose so an, is actually a
column vector. Any non zero elements in an, amn≠0∈an, is
given by (8). Here we use both subscript m and n for the
element index of input vector and the iteration (and or data
persentation) index, respectively. The corresponding set of
adjustable weights is also in the column form as
[
w n = w1n
]
w2 n L w pn . (10)
T
Given an input vector an and a weight vector wn, a single
output of ALC on the nth iteration is given by
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
yn = a n w n =
T
g
∑a
jn
85
w jn .
(11)
j =1
At the first right of equal sign in (11), the overall elements of
the weighted input including zero elements is summed up,
while in the second one, only g non zero elements to be
summed to get the same output, yn.
C. Learning Scheme and Generalization
The multi input-single output mapping of ALC is
determined by the stored weights. The weights are adjusted
based on the least mean square algorithm so that the
mapping achieves the desired function given a priori. Let the
output of ALC for the input vector a be y, the corresponding
training datum be h, and the output error between y and h be
ε. Given an input vector an and the training datum be hn, the
weights wjn (j=1,2,…,g) associated with the input are
adjusted according to
w j ( n+1) = w jn + δ jn ;
j = 1,2, L , g
(a)
(12)
where
δ jn = 2µε n a jn
ε n = hn − y n
The parameter µ is a gain factor that governs a learning
speed. Let the output error for the input an after the updating
be ε+. Then we have
ε n = hn − a Tn (w n + 2µε n a n )
+
= hn − yn − 2µε n a Tn a n
= (1 − 2µa Tn a n )ε n .
To make ε n+ ≤ ε n , the parameter µ is selected so it is in the
interval
0<µ ≤
1
.
T
2a n a n
(13)
The significant feature of CMAC is that the learning
scheme changes the output values for the nearby inputs,
because each of subregions that formed by amn∈an covers g2
inputs, ik (i=1,2). Therefore, similar inputs lead to similar
outputs even for untrained inputs. This property is called
generalization, which is of great use in the CMAC-based
modelling. Moreover, we can control the degree of
generalization by changing the size of g. A larger g gives
larger subregion and wider generalization region. This is to
be the reason so g is called a generalization parameter.
Fig. 2(a) shows the generalization region of CMAC with
g=4. We shall give a training signal to the input h(5,6). The
input then specifies four subregions denoted by the squares,
and four weights stored in the subregions are updated. The
number of updated weights for each input is represented by
“1”, “2”, “3”, and “4”. The outputs indicated by “4”, “3”,
(b)
Figrue 2. Generalization property of CMAC. (a) g=4. (b) g=2.
“2”, and “1” are increased by (2µε-aj), (3µε-aj/2), (µε-aj), and
(µε-aj/2), respectively, where ε- is the output error for the
input (5,6) before the updating and aj is the corresponding
association levels with respect to the covered inputs (1k,2k).
We see that the CMAC outputs are changed even for
untrained inputs. For comparisson, Fig. 2(b) shows the
generalization region of a CMAC with g=2. In this case, two
weights are updated, and the outputs indicated by “2” and
“1” are increased by (2µε-aj) and (µε-aj), respectively. We
see that the generalization region becomes smaller than the
previous one.
III.
CMAC-BASED MODELING
Fig. 3 shows a structure of the proposed modelling
system. Here, an input pair (x,y) is quantized by an interval
quantizer to get a quantized quantity pair, (1k,2k). The input
mapping uses both 1k and 2k to generate the association
address, m. Those quantized quantity pair is also utilizated to
calculate the association levels in m address, am. ALC then
uses the association levels am and an output h to be its
training data pair. After a set of weights adjusted in the
training session of ALC, the association levels weighted by a
fixed set of weights is summed up to get an approximation
~
of the desired output, h . Because of using a quantized
quantity pair in the input mapping, the dequantizer then
x and
needs to dequantize them to get the quantized input, ~
~y , that corresponds to the output h~ in h~ (~
~
x, y ).
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
86
Proceedings of CITEE, August 4, 2009
IV.
Figure 3. Generalization property of CMAC. (a) g=4. (b) g=2.
The gravimetric geoid data can be represented in general
form as h(x,y) where the spatial position, x and y, is referred
to longitude and latitude, respectively, and the gravimetric
geoid, h, is the geoid height at position (x,y). Given a set of L
positions, {x,y}∈RL, where x in [xmin,xmax] and y in
[ymin,ymax]. Let the selected ROI (region of interest) for the
dimension of x and y be in interval [1u0,1uM] and [1v0,1vN]
respectively so the ROI covers all position in datasets. By
deciding a number of 1Q and 2Q quantization levels for each
of ROI intervals, quantizer (1) will give a set of L positive
integer {1k,2k} where 1k in [1,1Q] and 2k in [1,2Q]. Equation
(3) then utilizated those quantized quantities, 1k and 2k, to
associate a number of g address vector as the element indexs
of a n ∈ R p (n=1,2,…,L). Those quantized quantities are also
used by (8) to calculate the corresponding association levels
in each associated address. ALC then uses both association
vector an and geoid height h be its input-output training. By
selecting a gain factor, µ, so it is in (13), a set of adjustable
weights, w ∈ R p , is trained by ALC using (12) until some
stoping criteria achieved. Given a fixed set of adjusted
weights, w*∈ R p , and an association vector, a n ∈ R p ,
~
n=1,2,…,L, an approximation of geoid height, hn , is
~
calculated by (11). The quantity hn is not only an
approximation value of true geoid height hn, but its
corresponding position, (~
x, ~
y ) , given by dequantizer (2) is
also an approximation value of true position (x,y).
The last statements in the previous paragraph confirms
the using of model assesment that incorporate the
~
approximated values in both h and (~
x , ~y ) . Reference [17]
was given a respective procedure in model assesment that
used by this paper. In that procedure, data and model in
geodetic coordinat system are transformed into geocentric
cartesian system and then distance of consequtive points in
the new coordinat system are calculated. Residual distances
between model and data are used to asses the model.
As shown in Fig. 2, generalization regions can be control
by using a pre-defined value of generalization parameter, g.
In “unpublished” [18], generalization g is pre-defined so it is
a positive integer in [1,max(1Q,2Q)). By first ploting the
distribution of TTG (tanda tinggi geodesi) to see its
coverage area and defining properly ROI so it covers all
TTG, then generalization g is approximately defined in such
so its value is equal or larger than the maximum distance
between the ROI frames and its respective TTG area frames.
For comparing purpose in this paper, we implement two
CMACs with different generalization parameter in the same
maximum quantization level of both x and y. Here we define
CMAC with g=73 and g=123 in the maximum quantization
level 1Q=2Q=256.
ISSN: 2085-6350
RESULT AND DISCUSSION
In this section, the distribution of TTG points around
Merapi and Merbabu volcanoes are first explained. Those
TTGs are then processed by the proposed method and the
results which are consist of contour lines of corresponding
geoid heights model given by two CMACs with different g,
wil be discussed in the second part of this section. Model
assesments are then presented in the last part and also closed
this section.
Fig. 4(a) and (b) show the distribution of TTG points and
its corresponding EGM96 and EIGEN-GL04C geoid height,
respectively. In geoid height plotting, difference of geoids by
its minimum value are used instead of its original one to
visually enhance its spatial distribution of geoid in each
TTG. In those figures we then know that TTGs which are lie
in the path from Surakarta to Yogyakarta, are mostly have
minimal geoid heights, while TTGs which are lained in the
path from around Muntilan to near Ambarawa are mostly
have maximal and uniform geoid heights. Different with
(a)
(b)
Figrue 4. Distribution of TTG and its corresponding geoid height, (a)
EGM96, (b) EIGEN-GL04C.
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
87
(a)
(a)
(b)
(b)
Figure 5. Contour lines of geoid model and difference between geoid
data and its model using CMAC with g=73, (a) EGM96, (b) EIGENGL04C
Figure 6. Contour lines of geoid model and difference between geoid
data and its model using CMAC with g=123, (a) EGM96, (b) EIGENGL04C.
TTGs in two paths explained previously, TTGs which are
lained in the path from Bawen to Boyolali are have
fluctuative geoid heights.
of extrapolating points to share their “knowledge” received
from respective trained points. Vertical lines that figured on
each TTGs in Fig. 5 and 6 are difference between geoid data
and its model. From those lines we visually know that
CMAC with smaller g has a smaller difference. This visually
assesment will be explored quantitatively in the next
discussion.
Fig. 5 and 6 show contour lines of geoid models which
are given by CMAC with g=73 and g=123, respectively. In
both figures, EGM96 model are placed in the up side, while
another one are in the bottom side. All figures use the same
resolution in contour lines calculating, that is, a number of
64 contour levels is defined on peak-to peak values of geoid
model. Those figures show that the geoid models given by
CMAC with smaller g has more peaks (top and bottom) than
others in geoid models given by CMAC with larger g.
CMAC with smaller g generalize a small area around trained
points and a narrow valley at near the volcanoes shown in
Fig. 5(a) and (b) is caused by the overlapping of
generalization regions is only in their edges. Different
situations as shown in Fig. 6(a) and (b) are given by CMAC
with larger g. In this situations, generalization regions which
almost covered all trained points are functioned as a bridge
Fig. 7 and 8 show the results of model assesment
procedure related to the geoid model displaying on Fig. 5
and 6, respectively. In that procedure, distance of
consequtive points in geocentric cartesian system are
calculated from both data and model. Residual distance
which are defined as difference distance between data and
model are then used to asses the model. As we have
expected, CMAC with smaller g has smaller residual
distance than other ones given by CMAC with larger g. For
CMAC with g=73, residual distance of EGM96 model is
about 0.00025 m in total absolut value and standard
deviation is about 7.47x10-6 m. These two values are lower
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350
88
Proceedings of CITEE, August 4, 2009
than another ones given by CMAC with g=123. This CMAC
produces EGM96 model with residual distance is about
0.00069 m and standard deviation is about 1.94x10-5 m. Fig.
7 and 8 are also showed that using the same CMAC, residual
distance that was given by EIGEN-GL04C models is larger
then it was given by EGM96 models.
(a) EGM96
(a) EGM96
(b) EIGEN-GL04C
Figure 8. Model assesment in using CMAC with g=123
REFERENCES
(b) EIGEN-GL04C
Figure 7. Model assesment in using CMAC with g=73
V.
[1]
[2]
CONCLUSION
In this paper, it is presented that CMAC was a method to
propose gravimetric geoid using EGM96 and EIGENGL04C geopotential developments. The advantages of the
proposed method are summarized as follows.
1) It is very well in local gravimetric geoid response.
[3]
[4]
2) It does not need to redesign when the datasets updated.
3) Spatial resolution is known and customized before the
development process, even if the datasets are limited.
CMAC is inherently capable of learning a multi-input multioutput non linear function. Using three input single output
CMAC’s, the proposed method for 2D non linear function
can straightforwardly be applied to 3D system such as
temporal variation of 2D gravity field. The extenxion will be
shown elsewhere.
[5]
[6]
[7]
[8]
[9]
ISSN: 2085-6350
J. S. Albus, “A new approach to manipulator control: The cerebellar
model articulation controller (CMAC),” Trans. ASME J. Dynam.
Syst., Meas., Contr., vol. 97, no. 8, pp. 220-227, Sept. 1975.
S. Jagannathan, “Discrete-time CMAC NN control of feedback
linearizable nonlinear systems under a persistence of excitation,”
IEEE Trans. Neural Netwroks, vol. 10, pp. 128-137, Jan. 1999.
Y. F. Peng, R. J. Wai, and C. M. Lin, “Implementation of LLCCresonant driving circuit and adaptive CMAC neural network control
for linear piezoelectric ceramic motor,” IEEE Trans. Industrial
Electronics, vol. 51, pp. 35-48, Feb. 2004.
R. J. Wai, C. M. Lin, and Y. F. Peng, “Adaptive hybrid control for
linear piezoelectric ceramic motor drive using diagonal recurrent
CMAC network,” IEEE Trans. Neural Netwroks, vol. 15, pp. 14911506, Nov. 2004.
C. S. Lin and H. Kim, “CMAC-based adaptive critic self-learning
control,” IEEE Trans. Neural Netwroks, vol. 2, pp. 530-533, Sept.
1991.
A. V. Sebald and J. Schlenzig, “Minimax design of neural net
controllers for highly uncertain plants,” IEEE Trans. Neural
Netwroks, vol. 5, pp. 73-82, Jan. 1994.
H. M. Lee, C. M. Chen, and Y. F. Lu, “A self-organizing HCMAC
neural-network classifier,” IEEE Trans. Neural Netwroks, vol. 14, pp.
15-27, Jan. 2003.
N. E. Cotter and O. N. Mian, “A pulsed neural network capable of
universal approximation,” IEEE Trans. Neural Netwroks, vol. 3, pp.
308-314, Mar. 1992.
Y. Iiguni, “Hierarchical image coding via cerebellar model arithmetic
computers,” IEEE Trans. Image Processing, vol. 5, pp. 1393-1401,
Oct. 1996.
Conference on Information Technology and Electrical Engineering (CITEE)
Proceedings of CITEE, August 4, 2009
[10] A. M. Iradat, A. Susanto, dan Litasari, Pemampatan data citra
menggunakan CMAC, Tesis S2, Jurusan Teknik Elektro, UGM,
2004.
[11] C. S. Lin and C. T. Chiang, “Learning convergence of CMAC
technique,” IEEE Trans. Neural Netwroks, vol. 8, pp. 1281-1292,
Nov. 1997.
[12] J. S. Ker, Y. H. Kuo, and B. D. Liu, “Hardware implementation of
CMAC neural network with reduced storage requirement,” IEEE
Trans. Neural Netwroks, vol. 8, pp. 1545-1556, Nov. 1997.
[13] F. J. G. Serrano, A. R. F. Vidal, and A. A. Rodriguez, “Generalizing
CMAC architecture and training,” IEEE Trans. Neural Netwroks, vol.
9, pp. 1509-1514, Nov. 1998.
[14] J. C. Jan and S. L. Hung, “High-order MS_CMAC neural network,”
IEEE Trans. Neural Netwroks, vol. 12, pp. 598-603, May 2001.
89
[15] Y. F. Wong and A. Sideris, “Learning convergence in the cerebellar
model articulation controller,” IEEE Trans. Neural Netwroks, vol. 3,
pp. 115-121, Jan. 1992.
[16] D. E. Thompson and S. Kwon, “Neighborhood sequential and
random training techniques for CMAC,” IEEE Trans. Neural
Netwroks, vol. 6, pp. 196-202, Jan. 1995.
[17] T. A. Sunantyo, Modeling of Local Geoid using Adaptive Scheme: A
Case Study around Merapi and Merbabu Volcanoes, Central Jawa,
Indonesia, Doctoral Dissertation, Faculty of Engineering, Gadjah
Mada University, 2008.
[18] A. M. Iradat, Analisis-Sintesis Citra Sekuensial Berdasar Watak Citra
melalui Implementasi CMAC sebagai Penyandi Laju Bit Rendah,
Laporan Hasil Penelitian untuk Disertasi, Program Studi Teknik
Elektro, Fakultas Teknik, Universitas Gadjah Mada, Mei, 2009.
Conference on Information Technology and Electrical Engineering (CITEE)
ISSN: 2085-6350