Full Paper

Transcription

Full Paper
ISSN 2394-3777 (Print)
ISSN 2394-3785 (Online)
Available online at www.ijartet.com
International Journal of Advanced Research Trends in Engineering and Technology (IJARTET)
Vol. II, Special Issue XXIII, March 2015 in association with
FRANCIS XAVIER ENGINEERING COLLEGE, TIRUNELVELI
DEPARTMENT OF ELECTRONICS AND COMMUNICATION ENGINEERING
INTERNATIONAL CONFERENCE ON RECENT ADVANCES IN COMMUNICATION SYSTEMS AND
TECHNOLOGIES
(ICRACST’15)
TH
25 MARCH 2015
SEGMENTATION OF OPTIC DISC AND OPTIC CUP USING SPATIALLY WEIGHTED FUZZY
C MEAN CLUSTERING AND SUPERPIXEL ALGORITHM
K.Gowri
ME-Applied Electronics
Shri Andal Alagar College of Engineering
Mamandur
Dr.T.R.Ganesh Babu
Professor/ECE
Shri Andal Alagar College of Engineering
Mamandur
Abstract- Glaucoma is one of the leading cause of blindness if
it is not detected and treated in proper way. When there is an
elevated intra ocular pressure from the normal condition, the
subject is affected by glaucoma and in this condition the
retinal nerve fiber layer and the optic disc are affected and this
leads to progressive loss of vision if not diagnosed and treated.
Detection of this Glaucoma is very difficult task and Current
tests using intraocular pressure (IOP) testing are not sensitive
enough for population based glaucoma screening. In this
paper, glaucoma screening is treated by optic disc & cup
segmentation. Spatially Weighted Fuzzy C Mean(SWFCM)
Clustering method is used to segment the optic disc and
Superpixel algorithm is used to segment the optic cup. The
segmented optic disc and cup are then used to compute the
CDR for glaucoma screening. optic nerve head is also called
as optic disc. optic disc and cup segmentation of fundus image
is considered for the detection of glaucoma.
Keywords:Fundus image, glaucoma , cup to disc ratio, super
pixel.
aqueous humor is present in the eye. The aqueous humor is
produced by the ciliary body and is drained through the Canal
of Schlemm. If the aqueous humor does not drain out
correctly, then pressure will build up in the eye. Figure 1(a)
shows the normal fluid flow, which indicates the normal
pressure that acts in the eye. Figure 1(b) shows the blocked
fluid which indicates the elevated pressure that acts in the eye.
This high pressure damages the optic nerve leading to
glaucoma. visual loss is prevented by the timely diagnosis and
referral for management of these diseases. Hence the shortage
of trained personnel leads to the need for automatic retinal
image analysis system. Glaucoma is the most common cause
of irreversible blindness in the world. The World Health
Organization estimated the number of people who became
blind from glaucoma is 4.4 million in 2002. Globally, 60.5
million people are affected by glaucoma in 2010.
I. INTRODUCTION
Glaucoma is one of the common causes of blindness. It causes
progressive degeneration of optic nerve fibers and leads to
structural changes of the optic nerve and a simultaneous
functional failure of the visual field. Since, glaucoma is
asymptomatic in the early stages and the associated vision loss
cannot be restored, its early detection and subsequent medical
treatment is essential to prevent further visual damage.
(a)Normal fluid view
(b) Blocked fluid flow
Figure 1: Glaucoma Illustration
A. Glaucoma
Glaucoma is a general term for a family of eye diseases,
which, in most cases, leads to increased pressure within the
eye and as a result, damages the optic nerve. It affects people
of all ages and initially marginal vision is lost. If proper
treatment for glaucoma is not taken, then the vision loss still
continues, leading to total blindness.A watery material called
B. Diagnosis
Screening for glaucoma is usually performed as the part of
standard eye examination performed by ophthalmologists. The
measurements used for the presence of glaucoma are the eye
pressure, size/and shape of the eye, anterior chamber angle,
cup to disc ratio, rim appearance and vascular changes.The
6
All Rights Reserved © 2015 IJARTET
ISSN 2394-3777 (Print)
ISSN 2394-3785 (Online)
Available online at www.ijartet.com
International Journal of Advanced Research Trends in Engineering and Technology (IJARTET)
Vol. II, Special Issue XXIII, March 2015 in association with
FRANCIS XAVIER ENGINEERING COLLEGE, TIRUNELVELI
DEPARTMENT OF ELECTRONICS AND COMMUNICATION ENGINEERING
INTERNATIONAL CONFERENCE ON RECENT ADVANCES IN COMMUNICATION SYSTEMS AND
TECHNOLOGIES
(ICRACST’15)
TH
25 MARCH 2015
fundus image morphological changes in optic disc, optic cup
imaging techniques is inevitable in this work of glaucoma
detection which is helpful for strengthening the glaucoma
examination.
C. System Overview
Fig 3 a) Green channel
The input color fundus image is captured by fundus camera.
The color fundus image refers to the inside back portion of the
eye. This image provides an important information of the
retina in eye. Blood vessels, optic disc are captured in a
fundus image. Optic Disc is used to describe the portion of
optic nerve.OD (optic disc) is the entry point for major blood
vessel that supply the retina.The optic nerve head is also
called the optic disc. OD (optic disc) is placed 3 to 4 mm to
the nasal side of fovea. Normally there is a small depression
seen at front of the optic disc, which is known as the optic cup
and its diameter is smaller than the diameter of the optic disc.
The figure 1 shows the input fundus image.
Optic Disc
Optic Cup
Fig 3 b) ROI image
II. SEGMENTATION OF OPTIC DISC USING
SWFCM ALGORITHM
Optic disc is segmented by using SWFCM method
(Keh –Shih Chuang et-al 2006).The Spatial Weighted Fuzzy C
Mean Clustering Algorithm is applied to the ROI image The
main drawbacks of the Fuzzy C Means clustering (FCM) is
very sensitive to noise, does not consider the spatial
information of pixels and in turn the segmentation result is
affected. To overcome the above drawbacks, SWFCM is
used.
One of the important characteristics of an image is that
neighboring pixels are highly correlated. The spatial
relationship is important in clustering, but it is not utilized in a
standard FCM algorithm. To exploit the spatial information, a
spatial function is defined as
Fig 1. Input Fundus Image
(1)
D. REGION OF INTEREST (ROI) IMAGE
The input fundus image has been taken by fundus
camera in RGB mode. Green plane (G) is considered for the
extraction of optic disc, as it provides better contrast than the
other two planes. The maximum brightest point in the (G)
plane is identified and considered . This brightest pixel will
always fall in the cup region. The approximate region of the
optic disc is to be selected around this identified brightest
point. After analyzing 100 number of fundus images of size
1504 x 1000 pixels, it is decided to consider a square of size
360 x 360 pixels with the brightest pixel as the centre point
and this region is found to cover mainly the entire optic disc
along with a small portion of other regions of the image.
Figure 3 a) shows the Green channel and figure 3 b) shows the
ROI Image.
Where represents a square window centered on pixel in
spatial domain. Larger window size may blur the images and
the lower window size does not remove the noise at high
density. Therefore, 5*5 window size is used in this work. Just
like the membership function, the spatial function represents
the probability that the pixels belong to the ith cluster. The
spatial function of pixel is large if the majority of its
neighborhood belongs to the same clusters. The spatial
function is incorporated into membership function as follows
(2)
Where p and q are controlling parameters of both functions.
The spatial functions simply strengthen the original
membership in a homogenous region, but it does not change
clustering result. However, this formula reduces the weighting
7
All Rights Reserved © 2015 IJARTET
ISSN 2394-3777 (Print)
ISSN 2394-3785 (Online)
Available online at www.ijartet.com
International Journal of Advanced Research Trends in Engineering and Technology (IJARTET)
Vol. II, Special Issue XXIII, March 2015 in association with
FRANCIS XAVIER ENGINEERING COLLEGE, TIRUNELVELI
DEPARTMENT OF ELECTRONICS AND COMMUNICATION ENGINEERING
INTERNATIONAL CONFERENCE ON RECENT ADVANCES IN COMMUNICATION SYSTEMS AND
TECHNOLOGIES
(ICRACST’15)
TH
25 MARCH 2015
of a noisy cluster in noisy pixels by the labels to its
neighboring pixels. As a result misclassified pixels from noisy
region or spurious blobs can easily be corrected. The spatial
FCM with parameter p and q is denoted If p = 1 and q = 0,
then SFCM1,0 is identical to the conventional FCM.
The clustering is a two-pass process at each
iteration. The first pass is the same as that in standard FCM to
calculate the membership function in the spectral domain. In
the second pass, the membership information of each pixel is
mapped to the spatial domain and the spatial domain function
is computed from that. The FCM iteration proceeds with the
new membership that is incorporated with spatial function.
The iteration is stopped when the maximum difference
between two cluster centers at two successive iterations is less
than 0.00001. After the convergence, defuzzification is
applied to assign each pixel to a specific cluster for which the
membership is maximal.
Algorithm
Step 1: Generate the random number with the range from 0
to1 to be the initial memberships. Let us consider the number
of cluster is N then calculate Vi using 3.
6
Where
The segmented image has three clusters, first and second
cluster represents the background, and third cluster represents
the optic disc. The clustered image is shown in the figure.
2.1.Detection of optic disc
The clustered image has three forms of cluster namely outer,
background, optic disc. To form the optic disc, initially
background cluster is eliminated by searching the corner
cluster index. From the two clusters, the optic disc is identified
by selecting the cluster index at the location of the brightest
point in the ROI image because the optic disc belongs to
higher intensity regions. Fig 4 shows the swfcm optic disc.
Outer
Back ground
Optic disc
Fig 4 SWFCM - optic disc
3
Step 2: Compute 4
2.2.Optic Disc Boundary Smoothening
4
Step 3: Map into the pixel position and calculate the modified
membership . Compute objective function J.
After extracting optic disc, elliptical fitting is applied for
smoothening the optic disc boundary (Fitzgibbon, A, Pilu, M
& Fisher, RB 1999). The labeling i.e., connected components
technique is applied to form the rectangle containing the
whole disc region as shown in Figure 5.a . The centroid of the
rectangle is taken as a center to draw an ellipse using the
equation (7) and (8) and that inscribed in the rectangle as
shown in figure 5.b. The area of the ellipse is calculated by
using the equation (9) (Ganeshbabu, TR & Shenbaga devi, S
2011).
Ellipse E
5
Step 4: Update the center 1
Step 5: Repeat steps 2 to step 4 until the following termination
criterion is satisfied:
X = a*(cosα cosβ) – b*(sinα sinβ)
7
Y = a*(cosα sinβ) + b*(sinα cosβ)
8
8
All Rights Reserved © 2015 IJARTET
ISSN 2394-3777 (Print)
ISSN 2394-3785 (Online)
Available online at www.ijartet.com
International Journal of Advanced Research Trends in Engineering and Technology (IJARTET)
Vol. II, Special Issue XXIII, March 2015 in association with
FRANCIS XAVIER ENGINEERING COLLEGE, TIRUNELVELI
DEPARTMENT OF ELECTRONICS AND COMMUNICATION ENGINEERING
INTERNATIONAL CONFERENCE ON RECENT ADVANCES IN COMMUNICATION SYSTEMS AND
TECHNOLOGIES
(ICRACST’15)
TH
25 MARCH 2015
Where ‘a’ is the major axis length (half of the rectangle width)
and b is the minor axis length (half of the rectangle height).
Here and α Varies from
2. A weighted distance measure combines color and spatial
proximity while simultaneously providing control over the
size and compactness of the super pixels.
The area of the ellipse is calculated by using the Equation
Algorithm:
Area=πab
9
Where
a = major axis length (half of the rectangle width) and
b = minor axis length (half of the rectangle height).
Thus the area of optic disc is calculated.
The connected component technique is applied to form the
rectangle containing the whole disc region as shown in the
figure 5a). Figure 5) b shows the contour of this ellipse being
shown on the original ROI.
Fig5 a) Elliptical optic disc
Fig 5 b) Imposed optic disc
III. SEGMENTATION OF OPTIC CUP
The optic cup is the inside portion of the optic disc. Optic
cup is segmented by using ROI Image. In this paper use the
simple linear iterative clustering algorithm (SLIC) to generate
superpixels by aggregating nearby pixels into superpixels in
images.
3.1.SLIC(Simple Linear Iterative Clustering)
SLIC algorithm is fast one, memory efficient and has great
boundary adherence. It is also simple to use.Simple linear
iterative clustering is an adaptation of k-means for super pixel
generation, with two important distinction:
1. The number of distance calculation in the optimization in
dramatically reduced by limiting the search space to a region
proportional to the super pixel size. This reduces the
complexity to be linear in the number of pixels N-and
independent of the no of super pixel k.
SLIC is simple to use and understand. By
default, the only parameter of the algorithm is k, the desired
number of approximately equally sized super pixels. For color
images in the CIELAB color space, the clustering procedure
begins with an initialization step where k initial cluster centers
Ci=[li ai xi yi]T are sampled on a regular grid spaced S pixels
apart. To produce roughly equally sized super pixels, the grid
interval is S= √N/K. The centers are moved towards the
lowest gradient position in a 3×3 neighborhood.
SLIC iteratively
Clustering is then applied. For each
searches for its best matching pixel from the 2S× 2
based on color and spatial proximity
neighborhood around
and then compute the new cluster center based on the found
pixel. The iteration continues until the distance between the
new centers and previous ones is small enough. Finally, a post
processing is applied to enforce connectivity.
Each pixel is associated with the nearest cluster
center whose search region overlaps its location. This is the
key to speeding up our Algorithm because limiting the size of
the search region significantly reduces the number of distance
calculations, and results in a significant speed advantage over
conventional k-means clustering where each pixel must be
compared with all cluster centers. This is the only possible
through the introduction of a distance measure D which
determines the nearest cluster center for each pixel .since the
expected spatial extent of a super pixel is a region of
approximate size S X S, the search for similar pixel is done in
a region 2S X 2S around the super pixel center.
Once each pixel has been associated to the nearest
cluster center, an update step adjusts the cluster centers to be
the mean [l a b x y]T vector of all the pixel s belonging to the
cluster .The L2 norm is used to compute a residual error E
between the new cluster center location and previous cluster
centre locations. Finally, a post processing steps enforces
connectivity by reassigning disjoint pixel to nearby super
pixels. Figure 6 shows the generation of super pixels using
fundus image.
9
All Rights Reserved © 2015 IJARTET
ISSN 2394-3777 (Print)
ISSN 2394-3785 (Online)
Available online at www.ijartet.com
International Journal of Advanced Research Trends in Engineering and Technology (IJARTET)
Vol. II, Special Issue XXIII, March 2015 in association with
FRANCIS XAVIER ENGINEERING COLLEGE, TIRUNELVELI
DEPARTMENT OF ELECTRONICS AND COMMUNICATION ENGINEERING
INTERNATIONAL CONFERENCE ON RECENT ADVANCES IN COMMUNICATION SYSTEMS AND
TECHNOLOGIES
(ICRACST’15)
TH
25 MARCH 2015
ds = sqrt(dx^2 + dy^2)
Spatial distance
which simplifies to the distance measure in equation 12 we
use in practice:
D=sqrt(dc2+(ds/S)2m2)
12
Fig 6: Superpixel Generation
3.2.Distance Measure
SLIC SUPER PIXELS correspond to clusters in the labxy
color-image plane space. This presents a problem in defining
the distance measure D,which may not be immediately
obvious. D computes the distance between a pixel I and
cluster center ck .A pixel’s color is represented in the CIELAB
color space [l a b]T , whose range of possible value is known
.The pixel’s position [x,y]T, on the other hand ,may take a
range of values that varies according to the size of the
image.Simply defining D to be the 5D Euclidean distance in
labxy Space will cause inconsistencies in clustering behavior
for different super pixel sizes. For large super pixel ,spatial
distances outweigh color proximity, giving more relative
importance to spatial proximity than color. This produces
compact super pixels that do not adhere well to image
boundaries.For smaller superpixels, the converse is true.
To combine the two distances dc and ds into a
single measure , it is necessary to normalize color proximity
and spatial proximity by their respective maximum distances
within cluster ,Ns and Nc. Distance (D’) is calculated by using
the equation 10
dc =√(lj-li)2+(aj-ai)2+(bj-bi)2,
ds=√(xj-xi)2+(yj-yi)2,
D’=√(dc/Nc)2+(ds/Ns)2
10
The maximum spatial distance expected with in a given
cluster
should correspond to the sampling interval,
Ns=S=√(N/K).To Determine the distance Nc is not so straight
forward,as color distance can vary significantly from cluster to
cluster and image to image.This problem can be avoided by
fixing Nc to a constant m. The m value is Substituting in the
equation 10,we get the equation 11
Distance = sqrt( dc^2 + (ds/s)^2*m^2 )
11
where:
dc = sqrt(dl^2 + da^2 + db^2) Colour distance
When m is small , the resulting super pixel s
adhere more tightly to image boundaries ,butr have less
regular size and shape.m is a weighting factor representing the
nominal maximum colour distance expected so that one can
rank colour similarity relative to distance similarity.try m for
L*a*b* space.After that it is very important to add features
Superpixel consists of a group of pixels with similar colors.
In this paper use center surround statistics (CSS)
from superpixels as a texture feature. To compute CSS, eight
spatial scale dyadic Gaussian pyramids are generated with a
ratio from 1:1 (level 0) to 1:256. Multiple scales are used as
the scale of the blob-like structures largely vary The dyadic
Gaussian pyramid is a hierarchy of low-pass filtered versions
of an image channel. It is accomplished by convolution with a
linearly separable Gaussian filter and decimation by a factor
of two. Then center surround operation between center (finer)
levels c =2,3,4 and surround levels (coarser) S = c + d with d
= 3,4 is applied to obtain six maps empirically computed at
levels of 2–5, 2–6, 3–6, 3–7, 4–7, and 4–8 from an image
channel.
Gaussian gradient will give change in the color
intensity in the targeted region with the neighboring area.From
the obtained features to determine the mean and variance.
3.3.3.SVM CLASSIFIER
In this paper use Support Vector Machine
(SVM) as a classifier (Smola et-al 1998). first create active
training data set to be a subset of the available training data set
(pool) and all training iteration is made on the active set. This
returns a preliminary classifier. Then classifier is used as a
pool evaluator training set is increased in every round of
training by misclassified in the previous iteration and continue
the process until there is no new action in the classification
accuracy or the maximum iterations have been reached. Here
it is not directly using the binary classification results from
SVM, decision function are used from SVM output values.
Each superpixel output is used as the decision
values for all pixels. After that smoothed decision value is
obtained by mean filter. Then a binary decision for all pixels
with a threshold is obtained using smoothed decision values.
Assign +1 and -1 to positive (CUP) and negative (NON-CUP)
samples, and the threshold is the average of them. Figure 6
10
All Rights Reserved © 2015 IJARTET
ISSN 2394-3777 (Print)
ISSN 2394-3785 (Online)
Available online at www.ijartet.com
International Journal of Advanced Research Trends in Engineering and Technology (IJARTET)
Vol. II, Special Issue XXIII, March 2015 in association with
FRANCIS XAVIER ENGINEERING COLLEGE, TIRUNELVELI
DEPARTMENT OF ELECTRONICS AND COMMUNICATION ENGINEERING
INTERNATIONAL CONFERENCE ON RECENT ADVANCES IN COMMUNICATION SYSTEMS AND
TECHNOLOGIES
(ICRACST’15)
TH
25 MARCH 2015
a,b,c shows the SLIC output , Segmented Optic cup With
Elliptical Fitting, Segmented Optic cup With Actual Fitting.
Fig 9 a)Normal image
CDR=0.321
Fig 7 a) slic oc
Fig 7 b)actual fitting
Fig 7 c)elliptical
fitting
IV. EXPERIMENTAL RESULTS
4.1.Optic Disc and Optic Cup Segmentation
The developed algorithm is tested on 10 normal
fundus images and 10 fundus images obtained from
glaucoma patients. Figures 8 (a) and 8 (b) shows the actual
segmented contour for optic cup and optic disc for normal and
glaucoma conditions.Figures 9 (a) and 9 (b) shows the
elliptical fitting optic cup and disc for normal and glaucoma
conditions respectively. The CDR values for normal and
abnormal images have been calculated by the developed
algorithm and they are listed in Tables 1 and 2. In the
Tables 1 and 2, the first column shows the subject number,
the second column indicates the CDR calculated by the
present algorithm with actual output, the third column
indicates the CDR calculated by the present algorithm with
elliptical output.Area of the resultant image is computed
interms of number of pixels and the cup to disc ratio is
computed as the ratio of the area of the optic cup to the
area of the optic disc.The CDR value greater than 0.3
which indicates glaucoma (Hossam El-Din MA Khalil et-al
2013).
Fig 8 a) Normal image
CDR=0.314
Fig 8 b) Abnormal image
CDR=0.404
Figure 8. CDR computations for both actual segmented
contour for optic cup and optic disc.
Fig 9 b) Abnormal image
CDR=0.416
Figure 9. CDR computations with elliptical fitting for both
optic cup and optic disc(k means clustering
TABLE 1.NORMAL
TABLE 2 ABNORMAL
Seria
CDR
CDR
l
Compute Compute
Num d from
d from
ber
actual Elliptical
of
image
image
Subj
1
0.3146
0.3210
Seria CDR
CDR
l
Compu Compute
Num
ted
d from
ber
from Elliptical
of
actual
image
1
0.4049
0.416
2
0.3141
0.3221
2
0.4040
0.4112
3
0.3147
0.3327
3
0.3792
0.3921
4
0.1815
0.2121
4
0.4639
0.4731
5
0.2736
0.2964
5
0.4901
0.5201
6
0.3141
0.3214
6
0.3741
0.3810
7
0.3251
0.3315
7
0.4925
0.5322
8
0.321
0.3354
8
0.4214
0.4314
9
0.2680
0.2816
9
0.450
0.4642
10
0.261
0.2898
10
0.4210
0.4321
V.CONCLUSION
An automated method for detection of glaucoma in retinal
fundus image is investigated. Such computerized system is
very useful for diagnosing glaucoma in the process of mass
screening for glaucoma. In this thesis, extracting features like
CDR from the fundus image have been proposed and used for
classification. To increase the accuracy further, more work
can be done on the following to enhance the accuracy in
detection of glaucoma.Analysis of ISNT ratio in fundus image
can be added as future work for strengthening the glaucoma
examination. Higher resolution fundus image may be used for
future work. As it provides clear visibility of optic disc, optic
11
All Rights Reserved © 2015 IJARTET
ISSN 2394-3777 (Print)
ISSN 2394-3785 (Online)
Available online at www.ijartet.com
International Journal of Advanced Research Trends in Engineering and Technology (IJARTET)
Vol. II, Special Issue XXIII, March 2015 in association with
FRANCIS XAVIER ENGINEERING COLLEGE, TIRUNELVELI
DEPARTMENT OF ELECTRONICS AND COMMUNICATION ENGINEERING
INTERNATIONAL CONFERENCE ON RECENT ADVANCES IN COMMUNICATION SYSTEMS AND
TECHNOLOGIES
(ICRACST’15)
TH
25 MARCH 2015
cup and blood vessels that increases the possibility of exact
segmentation of optic disc, optic cup and ISNT feature
calculation.
REFERENCES
[1]. Chih-Yin Ho, Tun-Wen Pai, Hao-Teng Chang & Hsin-Yi
Chen, 2011, ‘An automatic fundus image analysis system for
clinical diagnosis of glaucoma’, International Conference on
Complex, Intelligent, and Software Intensive Systems, pp.
559-564.
[2]. Chisako Muramatsu, Toshiaki Nakagawa, Akira Sawada,
Yuji Hatanaka, Tetsuya Yamamoto, Hiroshi Fujita, 2011,
‘Automated determination of cup-to-disc ratio for
classification of glaucomatous and normal eyes on stereo
retinal fundus images’, Journal of Biomedical optics, vol. 16,
no. 9, pp. 1-7.Technology (IJCCT), vol. 2, no. 6, pp. 7-10.
[3]. Fitzgibbon, A, Pilu, M & Fisher, RB 1999, ‘Direct least
square fitting of ellipses’, IEEE T Pattern, Anal, vol. 21, pp.
476-480.
[4]. Ganeshbabu, TR & Shenbaga devi, S 2011, ‘Automatic
detection of Glaucoma using fundus image’, European Journal
of Scientific and Research, vol. 59, no. 1, pp. 22-32.
[5]. Hossam El-Din MA Khalil, Mohamed Yasser Sayed Saif,
Mohamed Osman Abd El-Khalek & Arsany Maker 2013,
‘Variations of Cup -to –Disc Ratio in age Group(18-40) years
old’, Research in Opthalmology, vol. 2, no. 1. pp. 4-9
[6]. Jagadish Nayak, Rajendra Acharya, P, Subbanna Bha,
NakulShetty & Teik-Cheng Lim 2009, ‘Automated Diagnosis
of Glaucoma Using Digital Fundus Images’, Journal of
Medical
Systems,
vol.
33,
no.
5,
pp. 337-346.
[7]. Jun Cheng, Jiang Liu, Yanwu Xu, Fengshou Yin, Damon
Wing, Kee Wong, Ngan-Meng Tan, Dacheng Tao, Ching-Yu
Cheng, Tin Aung & Tien Yin Wong 2013, ‘Superpixel
Classification Based Optic Disc and Optic Cup Segmentation
for Glaucoma Screening’, IEEE Transactions on Medical
Imaging, vol. 32, no. 6, pp. 1019-1032.
[8]. Keh –Shih Chuang, Hong –Long Tzeng, SharonChen,
JayWu & Tzong –JerChen 2006, ‘Fuzzy C-Means Clustering
with Spatial Information For Image Segmentation’, Journal of
Computerized Medical Imaging and Graphics ,vol. 30, pp. 915.
[9]. Kevin noronha, Nayak, J&Bhat, SN 2006, ‘Enhancement
of retinal fundus Image to highlight the features for detection
of abnormal eyes’, IEEE Region 10 Conference TENCON,
pp. 1-4.
[10]. Liu, J, Wong, DWK, Lim, JH, Jia, X, Yin, F, Li, H,
Xiong, W & Wong, TY 2008, ‘Optic Cup and Disk Extraction
from Retinal Fundus Images for Determination of Cup-to-Disc
Ratio’, 3rd IEEE Conference on Industrial Electronics and
Applications, ICIEA, pp. 1828-18.
[11]. Madhusudan Mishra, Malaya Kumar Nath &
Samarendra Dandapat, 2011, ‘Glaucoma Detection from Color
Fundus Images’, International Journal of Computer &
Communication
[12]. Smola, Alex, J, Bernhard Schölkopf & Klaus-Robert
Müller, 1998, ‘The connection between regularization
operators and support vector kernels’, Neural Networks
vol.11, no. 4, pp. 637-649, 1998.
[13]. Wong, DWK, Liu, JH & Tan, NM 2009, ‘Intelligent
fusion of cup to disc ratio determination methods for
glaucoma detection in ARGALI’ , IEEE Annual International
Conference on Engineering in Medicine and Biology Society,
pp. 5777-5780.
[14]. Yuji Hatanaka, Atsushi Noudo, Chisako Muramatsu,
Akira Sawada, Takeshi Hara, Tetsuya Yamamoto & Hiroshi
Fujita 2012, ‘Vertical cup-to-disc ratio measurement for
diagnosis of glaucoma on fundus images’, Medical Imaging,
Proc. of SPIE, vol. 7624, pp. 76243C-(1-8).
[15]. Zhuo Zhang, Jiang Liu, Wing Kee Wong, Ngan, Meng
Tan, Joo Hwee Lim, Shijian Lu & Huiqi Li, 2009, ‘Convex
Hull Based Neuro-Retinal Optic Cup Ellipse Optimization in
Glaucoma Diagnosis’, 31st Annual International Conference
of the IEEE ,EMBS Minneapolis, Minnesota, USA, pp. 14411444.
12
All Rights Reserved © 2015 IJARTET