performance evaluation of brain tumor diagnosis techniques in mri
Transcription
performance evaluation of brain tumor diagnosis techniques in mri
International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 PERFORMANCE EVALUATION OF BRAIN TUMOR DIAGNOSIS TECHNIQUES IN MRI IMAGES V.Kala1, Dr.K.Kavitha2 1 M.Phil Scholar, Computer Science, Mother Teresa women’s university, (India) 2 Assistant Professor, Dept. of. Computer Science, Mother Teresa women’s university, (India) ABSTRACT Magnetic Resonance Imaging (MRI) images are generally employed in ischemic stroke diagnosis since it is quicker accomplishment and compatibility with majority life affirms devices. This paper proposed an efficient approach for automated ischemic stroke detection of employing segmentation, feature extraction, median filtering and classification that distinguish the region of ischemic stroke from sizeable tissues in MR images. The proposed approach comprises of five stages such as pre-processing, segmentation, median filtering, features extraction and classification. The former ischemic stroke detection is presented to enhance accuracy and efficiency of clinical pattern. The experimental results are numerically evaluated by a human proficient. The average overlap utility, average accuracy and average retrieve between the results found employing our proposed scheme. Key Words: Classification And Ischemic Stroke, Feature Extraction, Median Filtering ,MRI Images, Segmentation I.INTRODUCTION Generally, a stroke is denoted as a Cerebrovascular Accident (CVA) which is the frequent loss of the function of brain owing to interference in the blood issue to the brain. Such can be because of the ischemia induced by blockage or a bleedings a solution, the impressed brain area cannot purpose that may effect in an unfitness to precede one or more branches on one slope of the human body, unfitness to realize or articulate speech, or an unfitness to assure one slope of the visual area. An ischemic stroke is sometimes addressed in a hospital with thrombolytic agent who is also known as ‘clot buster’ and approximately hemorrhagic strokes gain from neurosurgery. Since the mean life span of human has enhanced, stroke gets turn the third extending campaign of death globally behind heart disease and also cancer. Hazard components for stroke admit geezer hood, cigarette smoking, eminent cholesterol, high blood pressure, former stroke or Transient Ischemic Attack (TIA), a test fibrillation, and diabetes. Our proposed scheme is an approach for the detection of stroke to settle the small area of ischemia for a distinct diagnosis and to assort whether the MR image is getting stroke or not. II. RELATED WORK The precision of the brain standardization approach immediately impacts the precision of statistical investigation of operational Magnetic Resonance Imaging (MRI) information. The medical secular lobe and cortical stratum structures necessitate an exact enrolment approach owing to prominent bury subject variance. Innovation of fully automated MRI post treating pipeline directed to minimize the error at the process of registration 1636 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 throughout group analyzes and we will establish their transcendence over two generally employed registration approaches by leading comprehensive surface to surface length quantifications throughout blunder cortical and sub cortical areas. Areas in 3-D Magnetic Resonance Images (MRI) of brain can be assorted utilizing protocols for manually sectioning and marking structures. For prominent cohorts, expertness and time essentials build such approach visionary. In order to attain mechanization, a single segmentation can be disseminated to some other single employing an anatomical reference symmetry approximation linking the atlas image to the objective image. The precision of the leading target marking has been determined but can possibly be enhanced by aggregating multiple segmentations employing decision fusion process. Though researches have furnished abundant manifest for caused advances in psychological and physiological welfare, trivial is recognized about potential links to brain structure of pattern. Applying high-resolution Magnetic Resonance Images of 22 Tai Chi Chuan (TCC) practicing and 18 assures checked for age, education and sex. We take off to analyze the fundamental anatomical correlatives of semi-permanent Tai chi pattern at two dissimilar levels of regional particularity. The structure of mean examples of anatomy, besides retrogression investigation of anatomical constructions is fundamental issues in medical field research, for example in the analysis of brain growth and disease procession. While the fundamental anatomical operation can be patterned by arguments in a Euclidian space, authoritative statistical approaches are applicable. Recent epoch work proposes that efforts to depict anatomical reference divergences employing flat Euclidian spaces counteract our power to constitute natural biological variance. All areas of neuroscience which utilize medical imaging of brain require for transmitting their solutions with address to anatomical areas. Particularly, relative morph metric and the group investigation of operational and physiologic data necessitate brains co-registration to demonstrate agreements throughout brain structures. It is considerably demonstrated that additive registration of one brain image to another image is unequal for adjusting brain structures, so legion algorithms induce issued to nonlinearly register brain images to each other. III. PROPOSED SYSTEM In our proposed approach we defeat the trouble and withdraw of existing approach. There are five stages are applied in our process. At feature extraction using GLCM and HOG (Histogram of Gradient) is employed and for assortment SVM and Neuro fuzzy ANFIS also presented. From the above process, we decide the region of ischemic stroke in MR images. MR images are more approachable, less costly and faster particularly in critically ill patients. By using our proposed approach we can obtain high accuracy images and the overall efficiency of the system is enhanced. 3.1. Data Flow Our proposed approach contains following four stages · Pre-processing · Segmentation · Feature extraction · Classification 1637 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 3.2. Pre-Processing In pre-processing approach median filters are employed to eliminate noise from the MRI input images. It is frequently suitable to be capable of executing few kind of noise diminution on images or signals. The median filters are known as nonlinear digital filters, frequently utilized to eliminate noise. Such noise elimination is a distinctive pre-processing level to enhance the solutions of more recent processing. Median filtering is widely utilized in digital image processing approach. 3.3. Architecture Diagram 3.4. Median Filtering As we have encountered that smoothing filters decrease noise. Nevertheless, the fundamental presumption is that the adjacent pixels represent extra samples of the like measures as the source pixel that is they constitute the same characteristic. At the image edges, this is obviously not true, and blurring of characteristics effects. We have employed convolution approach to enforce weighting kernels as a locality function that presented a linear 1638 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 procedure. There are nonlinear locality functions which can be executed for the intention of noise removal which can execute a better task of maintaining edges than Simple Smoothing Filters. 3.5. Segmentation Here, we really extract impressed region from the input image which a part of it and that comprises exactly the postcode. The aim of segmentation is to vary and simplify the cooperation of an input image into something which is more significant and lighter to analyze. 3.6. Feature Extraction (GLCM) Here, we are going to extract the video feature by GLCM and a gray level coincidence matrix (GLCM) comprises information concerning the situations of pixels causing similar gray level measures. Then, we compute various movement features at each and every point with local secular units separated in order to regard straight of motions. We compute the fluctuation between the each and every frame. Such measures will be employed as feature measures of video. The GLCM is determined by, Where ‘nij’ is defined as the number of occurrences and that possess the pixel values ‘(i,j)’ resting at the distance ‘d’ in the input image. The above co-occurrence matrix ‘Pd’ contains the dimension of about ‘n× n’, where ‘n’ is denoted as the number of gray levels in the input image. 3.7. Classification In this section, we are going to classify the input image whether the image is frontal or non-frontal image by employing Support Vector Machine (SVM) classifier. SVMs are also known as support vector networks that are monitored discovering examples with related discovering algorithms which analyze information and distinguish patterns, employed for the regression analysis and classification process. 3.8. Svm Classifier · Data setup: our proposed dataset comprises three categories, each ‘N’ samples. The information is‘2D’ plot source information for visual review. · SVM with analogy kernel (-t 0) and we require to discover the better parameter measure C employing 2fold cross establishment. · After detecting the better parameter measure for C, we aim the full data again employing such parameter measure. · Plot support vectors · Plot decision area SVM functions input vectors to a more eminent dimensional space vector where an optimum hyper plane is fabricated. Among the various hyper planes uncommitted, there is only too hyper plane which increases the length between them self and the closest data vectors of each and every class. Such hyper plane that increases the margin is known as the optimal distinguishing hyper plane. The margin is determined as the addition of hyper plane distances to the nearest training vectors of each and every category. 1639 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 IV. EXPERIMENTAL RESULTS Histogram of Oriented Gradients appropriates edge or gradient constructions which are feature of local shape. HOG is an image descriptor which is established on the image’s gradient preferences. Here we extract the mathematical measure from HOG only; HOG descriptor is established on dominant image edge orientations and the image are splitted into cells. 1640 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 V. CONCLUSION A Computer Aided Detection (CAD) system is capable of describing few ischemic stroke has been formulated. We have formulated an automated approach for the ischemic stroke detection in brain of MR images employing segmentation, feature extraction and classification. Here we enhance the precision by applying classification and by using SVM to discover the stoke level in the brain image but which is deficient in detecting precision, rather this we may employ some other progress classifier like GMM, HMM, Feed Forward Neural Network and thus the accuracy of the input image can be enhanced. REFERENCE [1]. Akansha Singh , Krishna Kant Singh, “A Study Of Image Segmentation Algorithms For Different Types Of Images”, International Journal of Computer Science Issues, vol. 7,Issue 5, pp 414-417,2010. [2]. Mansur Rozmin, Prof. Chhaya Suratwala, Prof. Vandana Shah,”Implementation Of Hard C-Means Clustering Algorithm For Medical Image Segmentation”, Journal Of Information Knowledge and Research in Electronics and Communication Engineering,vol.2, no.2, pp 436-440,Nov12-Oct13. [3]. Rajesh Kumar Rai, Trimbak R. Sontakke, “Implementation of Image Denoising using Thresholding Techniques”,International Journal of Computer Technology and Electronics Engineering (IJCTEE),vol.1,no. 2, pp 6-10. [4]. T.Kalaiselvi, S.Vijayalakshmi, K.Somasundara, “Segmentation of Brain Portion from MRI of Head Scans Using Kmeans Cluster”, International Journal of Computational Intelligence and Informatics ,vol. 1,no. 1, pp 75-79,2011. [5]. S.S Mankikar , “A Novel Hybrid Approach Using Kmeans Clustering and Threshold filter For Brain Tumor Detection”, International Journal of Computer Trends and Technology, vol. 4, no.3, pp 206-209,2013. [6]. M.C. Jobin Christ, R.M.S.Parvathi, “Segmentation of Medical Image using Clustering and Watershed Algorithms”, American Journal of Applied Sciences, vol. 8, pp 1349-1352, 2011. [7]. Manali Patil, Mrs.Prachi Kshirsagar, Samata Prabhu, Sonal Patil, Sunilka Patil,” Brain Tumor Identification Using K-Means Clustering”, International Journal of Engineering Trends and Technology,vol. 4,no. 3,pp 354-357,2013. [8]. P.Dhanalakshmi , T.Kanimozhi, “Automatic Segmentation of Brain Tumor using K-Means Clustering and its Area Calculation”, International Journal of Advanced Electrical and Electronics Engineering ,vol. 2,no. 2,pp 130-134,2013. [9]. Sanjay Kumar Dubey, Soumi Ghosh, “Comparative Analysis of K-Means and Fuzzy C Means Algorithms”, International Journal of Advanced Computer Science and Applications, vol. 4, no. 4, pp 35-39, 2013. [10]. M. Masroor Ahmed, Dzulkifli Bin Mohamad, “Segmentation of Brain MR Images for Tumor Extraction by Combining Kmeans Clustering and Perona-Malik Anisotropic Diffusion Model”, International Journal of Image Processing, vol. 2 , no. 1, pp 27-34,2008. [11]. Anam Mustaqeem, Ali Javed, Tehseen Fatima, “An Efficient Brain Tumor Detection Algorithm Using Watershed & Thresholding Based Segmentation”, I.J. Image, Graphics and Signal Processing, vol. 10,no. 5, pp 34-39,2012. 1641 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 [12]. P.Vasuda, S.Satheesh, “Improved Fuzzy C-Means Algorithm for MR Brain Image Segmentation”, International Journal on Computer Science and Engineering (IJCSE), vol. 02, no.05, pp 1713-1715, 2010. [13]. Ananda Resmi S, Tessamma Thomas, ”Automatic Segmentation Framework for Primary Tumors from Brain MRIs Using Morphological Filtering Techniques”, in 5th Int Conf on Biomedical Engineering and Informatics,2012,IEEE. 1642 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 A REVIEW OF BENCHMARKING IN SERVICE INDUSTRIES Sunil Mehra1, Bhupender Singh2, Vikram Singh3 1 M.Tech. Student, 2Asst. Professor, 3Associate Professor, Mechanical Engineering, YMCA University of Science and Technology, (India) ABSTRACT Benchmarking is recognized as a fundamental tool for continuous improvement of quality. The objective of this paper is to examine service factor that contribute to the effectiveness of benchmarking in service industries. This study found that complexity and flexibility has significant correlation with effectiveness of benchmarking in service industries. The purpose of this paper is to understand of success factors aimed at increasing service revenue in manufacturing companies or service industries. We find out different types of success factors and those factors are improving performance of any type of service industries. Keywords: Benchmarking, Finding Factors, Implementing Benchmarking, Organisation, Service I. INTRODUCTION Benchmarking is one of the most useful tools of transferring knowledge and improvement into organisations as well as industries (Spendolini, 1992; Czuchry et al., 1995). Benchmarking is used to compare performance with other organisations and other sectors. This is possible because many business processes are basically the same from sector to sector. Benchmarking focused on the improvement of any given business process by exploiting best rather than simply measuring the best performance. Best practices are the cause of best performance. Companies studying best practices have the greatest opportunity for gaining strategic, operational, and financial advantages. Benchmarking of business processes is usually done with top performing companies in other industry sectors. The systematic discipline of benchmarking is focused on identifying, studying, analysing, and adapting best practices and implementing the results. The benchmarking process involves comparing one’s firm performance on a set of measurable parameters of strategic importance against those firms known to have achieved best performance on those indicators. Development of benchmarking is an iterative and ongoing process that is likely to involve sharing information with other organisations working with them towards a satisfying metrology. Benchmarking should be looked upon as a tool for improvement within a wider scope of customer focused improvement activities and should be driven by customer and internal organisation needs. Benchmarking is the practices of being humble enough to admit that someone else is better at something and wise enough to learn how to match even surpass them at it. 1643 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 II. IMPLEMENTATION OF BENCHMARKING There are five phases for implementation of benchmarking 2.1 Planning Planning is the first step of implementation of benchmarking in any organisation. During this phase the organisation determines which process to benchmark and against what type of organisation. 2.2 Analysis During this phase to analysis is performed for the performance gap between the source organisation and the receiver organisation. 2.3 Integration It involves the preparation of the receiver for implementation of actions. 2.4 Action This is the phase where the actions are implemented within the receiver organization. 2.5 Maturity This involves continuous monitoring of the process and enables continuous learning and provides input for continuous improvement within the receiver organization. III. SERVICE The concept of service can be described as the transformation of value, an indescribable product, from the service supplier (also termed the provider) to the customer (also termed the consumer). The process of transformation can be set in motion by a customer whose needs can be provided by the supplier, by a service supplier who offers a particular service to the customer. The definitions usually include: 1644 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 · Public transportation. · Public utilities (telephone communication, energy service, sanitation stores). · Restaurants, hotels and motels. · Marketing (retail food, apparel, automotive, wholesale trade, department stores). · Finance (commercial banks, insurance, sales finance, investment). · Personal service (amusements, laundry and cleaning, barber and beauty shops). · Professional services (physicians, lawyers). · Government (defense, health, education, welfare, municipal services). · News media IV. LITERATURE REVIEW 4.1 Roger Moser et al., (2011), In order to develop a benchmarking framework for supply network configuration we draw upon insights various theories addressing different levels: the dyadic relationship, the supply chain, and the network levels. The all three levels for the benchmarking of supply network configuration. The foundation for the development of our benchmarking framework for supply network configuration is primarily based on the theories of relationships and networks. 4.2 Heiko Gebauer et al., (2011), The different requirements of the service strategies described as “after-sales service providers” and “customer support service providers” influence the logistics of spare parts. The service organisations would face increasing pressure to improve their financial performance, and compared with the corresponding values for the manufacture and distribution of the finished product within the company. 4.3 Min et al., (Hokey 2011) The benchmarking process begins with the establishment of service standards through identification of service attributes that comprise service standards. Since serving customer better is the ultimate goal of benchmarking. 4.5 Panchapakesan padma et al., (2010) Several researchers have established that service quality not only influences the satisfaction of buyers but also their purchase intentions. Even though there are other antecedents to customer satisfaction , namely, price, situation, and personality of the buyer, service quality receives special attention from the service marketers because it is within the control of the service provider and by improving service quality its consequence customer satisfaction could be improved, which may be turn influence the buyer’s intention to purchase the service. 4.5 Luiz Felipe Scavard et al., (2009) Product variety proliferation is a trend in many industry sectors worldwide. Benchmarking should be a reference or measurement standard for comparison a performance measurement that is the standard of excellence for a specific business and a measurable, best in class achievement. Benchmarking as the continuous process of 1645 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 measuring products, services, and practices against the toughest competitors or those organization known as industry leaders. 4.6 Monika koller, Thomas Salzberger, (2009) Benchmarking is a tool which is widely used in both the manufacturing as well as the service industry. When we compared to the manufacturing sector, benchmarking in the service sector is more difficult because idiosyncrasies of particular fields of application. Bhutta and Huq (1999) gives that benchmarking is not only a comparative analysis, it is a process to establish the ground for creative breakthroughs and a way to move away from tradition. 4.7 Okke Braadbaart, (2007), Benchmarking has significant potential for application in the public sector. According to Magd and Curry (2003), benchmarking is a powerful vehicle for quality improvement and a paradigm for effectively managing the transformation of public-sector organizations into public-sector organizations of quality. The collaborative benchmarking literature focuses on information sharing and the positive effect this has on the quality and quantity of information about what public sector organisations do and how well they do it. 4.8 Sameer Prasad, Tata, (2006) Employing an appropriate theory building process can help us improve our precision and understanding. Both precision and understanding are important element of benchmarking. The greater the degree of precision allows for a finer ability to make comparisons. Greater power provides an understanding on the way to improve upon the service. 4.9 Nancy M. Levenburg, (2006) Benchmarking is the process of comparing one’s practices and procedures against those believed to be the best in the industry. While benchmarking efforts focused on manufacturing and logistics, the process has grown to encompass a wider array of activities including exporting, quality goals in service systems, supply chain interface, employee practices and brand management. This study aims to gain insight into practice that can enable organizations to utilize the internet more effectively for customer service purpose. 4.10 Ashok Kumar et al., (2006) Benchmarking has been variously defined as the process of identifying, understanding, and adapting outstanding practices from organizations anywhere in the world to help your organizations improve its performance. It is an activity that looks outward to find best practice and high performance and then measures actual business operations against those goals. V. FINDING FACTORS With the help of this literature review we find out different types of success factors and those factors are improving performance of any type of service industries. Many service factors are finding in service industries. Some factors are also inter relate to each other. Service factor are:- 1646 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 Sr. ISBN: 978-81-931039-3-7 Factor Author Define Unique or superior Melton and Hartline, Offering superior core attributes and supporting service 2010 services. Synergy Ottenbacher No. 1 2 and Capabilities Harrington, 2010 3 customer Carbonell et al., 2009 involvement Operational outcomes and innovation volume, but no impact on competitive superiority and sales performance. 4 Technology Lin et al., 2010 By applying more advanced marketing information systems based on the data acquired from their customers, companies are able to create more service innovations to explore potential markets 5 6 Knowledge Leiponen, Collective ownership of knowledge should be management 2006 promoted. Culture Liu (2009) Supportive culture as a construct of complementary dimensions consisting of innovative 7 Market orientation Atuahene-Gima (1996) The organization-wide collection and dissemination of market information, as well as the organizational responsiveness to that information. 8 Process quality Avlonitis et al., 2001 This factor is important in all phases from idea generation and analysis to concept development, testing, and launch. 9 Cross-functional Storey et al.,2010 It seems to be important in companies which rely involvement heavily on tacit knowledge, where the codification of information is difficult 10 employee Ordanini involvement Parasuraman, 2011, and Extensive internal marketing is conducted to raise support and enthusiasm for the product VI. CONCLUSION From this review it is seen that Benchmarking is an important tool to improve quality of the product/service in manufacturing as well as service industries. The main objective of the review to highlight the main factor in service industries and these are interrelate each other. These different types of successive factor are improving 1647 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 performance of any type service industries as well as manufacturing and those factors are also useful for implementation of benchmarking. A service entails a unique experience between the service provider and service customer. The constellation of features and characteristics inherent in a service offering takes place during its development. Hence, it is important to be aware of certain elements which contribute to the success of a service while designing it. Benchmarking establishes company’s true position versus the rest, making thus easier for the company to raise organizational energy for change and develop plans for action. REFERENCES [1] Heiko Gebauer, Gunther Kucza, Chunzhi Wang, (2011),"Spare parts logistics for the Chinese market", Benchmarking: An International Journal, Vol. 18 Iss: 6 pp. 748 – 768. [2] Roger Moser, Daniel Kern, Sina Wohlfarth, Evi Hartmann, (2011),"Supply network configuration benchmarking: Framework development and application in the Indian automotive industry", Benchmarking: An International Journal, Vol. 18 Iss: 6 pp. 783 – 801. [3] Sameer Prasad, Jasmine Tata, (2006),"A framework for information services: benchmarking for countries and companies", Benchmarking: An International Journal, Vol. 13 Iss: 3 pp. 311 – 323. [4] Nancy M. Levenburg, (2006),"Benchmarking customer service on the internet: best practices from family businesses", Benchmarking: An International Journal, Vol. 13 Iss: 3 pp. 355 – 373 [5] Monika Koller, Thomas Salzberger, (2009),"Benchmarking in service marketing - a longitudinal analysis of the customer", Benchmarking: An International Journal, Vol. 16 Iss: 3 pp. 401 – 414. [6] Okke Braadbaart, (2007),"Collaborative benchmarking, transparency and performance: Evidence from The Netherlands water supply industry", Benchmarking: An International Journal, Vol. 14 Iss: 6 pp. 677 – 692. [7] Ashok Kumar, Jiju Antony, Tej S. Dhakar, (2006),"Integrating quality function deployment and benchmarking to achieve greater profitability", Benchmarking: An International Journal, Vol. 13 Iss: 3 pp. 290 – 310. for achieving service innovation. Decision Sciences, 40, 431-475. [8] Luiz Felipe Scavarda, Jens Schaffer, Annibal José Scavarda, Augusto da Cunha Reis, Heinrich Schleich, (2009),"Product variety: an auto industry analysis and a benchmarking study", Benchmarking: An International Journal, Vol. 16 Iss: 3 pp. 387 – 400. [9] Panchapakesan Padma, Chandrasekharan Rajendran, Prakash Sai Lokachari, (2010),"Service quality and its impact on customer satisfaction in Indian hospitals: Perspectives of patients and their attendants", Benchmarking: An International Journal, Vol. 17 Iss: 6 pp. 807 – 841 [10] Hokey Min, Hyesung Min, (2011),"Benchmarking the service quality of fast-food restaurant franchises in the USA: A longitudinal study", Benchmarking: An International Journal, Vol. 18 Iss: 2 pp. 282 – 300. [11] G.M. Rynja, D.C. Moy, (2006),"Laboratory service evaluation: laboratory product model and the supply chain", Benchmarking: An International Journal, Vol. 13 Iss: 3 pp. 324 – 336. [12] Heiko Gebauer, Thomas Friedli, Elgar Fleisch, (2006),"Success factors for achieving high service revenues in manufacturing companies", Benchmarking: An International Journal, Vol. 13 Iss: 3 pp. 374 – 386. [13] Jo Ann M. Duffy, James A. Fitzsimmons, Nikhil Jain, (2006),"Identifying and studying "best-performing" services: An application of DEA to long-term care", Benchmarking: An International Journal, Vol. 13 Iss: 3 pp. 232 – 251. 1648 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 [14] Li-Jen Jessica Hwang, Andrew Lockwood, (2006),"Understanding the challenges of implementing best practices in hospitality and tourism SMEs", Benchmarking: An International Journal, Vol. 13 Iss: 3 pp. 337 – 354. [15] Lin Peter Wei-Shong, Mei Albert Kuo-Chung, (2006),"The internal performance measures of bank lending: a value-added approach", Benchmarking: An International Journal, Vol. 13 Iss: 3 pp. 272 – 289. [16] Mahmoud M. Nourayi, (2006),"Profitability in professional sports and benchmarking: the case of NBA franchises", Benchmarking: An International Journal, Vol. 13 Iss: 3 pp. 252– 271. [17] Sarath Delpachitra, (2008),"Activity-based costing and process benchmarking: An application to general insurance", Benchmarking: An International Journal, Vol. 15 Iss: 2 pp. 137 – 147. [18] Hervani, A.A., Helms, M.M. and Sarkis, J. (2005), “Performance measurement for green supply chain management”, Benchmarking: An International Journal, Vol. 12 No. 4, pp. 330-53. [19] Kyro¨, P. (2003), “Revising the concept of benchmarking”, Benchmarking: An International Journal, Vol. 10 No. 3, pp. 210-25. [20] Simatupang, T.M. and Sridharan, R. (2004), “Benchmarking supply chain collaboration, an empirical study”, Benchmarking: An International Journal, Vol. 11 No. 5, pp. 484-503. 1649 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 IDENTIFICATION OF BARRIERS AFFECTING IMPLEMENTATION OF 5S Sunil Mehra1, Rajesh attri2, Bhupender Singh3 1 M.Tech. Student, 2,3Asst. Professor, Mechanical Engineering, YMCA University of Science and Technology, (India) ABSTRACT The 5S methodology is the best tool for generating a change in attitude among employee and serves as a way to engage important activities the workplace. But, implementation of this universally accepted and challenging system is not an easy task by any means as it requires establishing new cultures, changing attitudes, creating good work environments. There are certain barriers which effects the implementation of 5S in the manufacturing organisation. The aim of the present work is to identify the different types of barriers which affect the implementation of 5S in manufacturing organisation. Keywords: 5s, Barriers, Implementation, Manufacturing Organisation I. INTRODUCTION The 5S methodology is a very suitable way to start the process of continuous improvement (Carmen Jaca et al., 2013). The 5S methodology is one of the best tools for generate a change in attitude among workers and serves as a way to engage improvement activities the workplace (Gapp et al., 2008). This methodology was developed in Japan by Hirano (1996). The 5S name corresponding to the first latter of five Japanese word – Seiri, Seiton, Seiso, Seiketsu, Shitsuke, (Ramos Alonso 2002) and their rough English equivalents – Sort, Set in order (Straighten), Shine, Standardize, Sustain. Hirano establishes that Lean Culture requires a change in people’s mentality as well as applying 5S as a requirement for the implementation of other actions to achieve improvement and as a basic step towards eliminating waste. In Japanese culture each word that makes up 5S means the following (Ramos Alonso 2002): 1. Seiri (Sort) – In which, we identify what is needed to do daily work, what is not needed, and what can be improved. 2. Seiton (Straighten/ Set in Order) – In which, we organise the work area with the best locations for the needed items. 3. Seiso (Shine) – In which, we Clean or remove reasons for unorganized, unproductive and unsafe work. Create measures and preventative maintenance to ensure the Shine step. 4. Seiketsu (Standardize) – In which, we provide procedures to ensure understanding of the process. This S supports the first 3 S’. Keep using best practices. 5. Shitsuke (Sustain) – In which, we set up the system to ensure the integrity of the process and build it so it that improvement is continuous. 1650 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 5S can ultimately be applied to any work area, in and outside manufacturing. The same techniques apply to any process including those in the office (Liker et al., 2004). The main objective of the current research is to identify the barriers in the implementation of 5S in Indian manufacturing organizations. II. BENEFITS OF 5S • 5S will improve work efficiency, safety and also improves productivity. • A clean and organised workplace is safer. It decreases the possibility of injuries occurring. • Increase product quality and process quality. • 5S will motivate and involve your employees. • 5S will organize your whole organisation as well as workplace. • 5S will remove clutter from workspace. III. IMPLEMENTATION OF 5S The 5S implementation requires commitment from both top management and everyone in the organisation. The 5S practice requires investment in time and if properly implemented it has a huge impact on organisational performance (Ho 1999a; linker 2004; linker and Hoseus 2008). The effective implementation of the 5S method is the responsibility of the management and the entire team of employees. The implementation should be carried out after prior training and making staff aware of the validity and the effectiveness of the method used. Literature review and experiences of managers and academicians reveal that implementation of 5S is not an easy task by means as it requires establishing new cultures, changing attitudes, creating good work environment’s and shifting the responsibility to the every employee of the organisation. The main purpose of implementing 5S is to achieve better quality and safety. IV. IDENTIFICATION OF BARRIERS 4.1 Lack of Top Management Commitment Top management is to control and help the continuous improvement activities, it relate the activities to business target and goals. Top management support and commitment is necessary for any strategic program success (Hamel & Prahalad, 1989; Zhu & Sarkis, 2007). Lack of commitment from top management is a chief barrier for successful adoption of green business practices (Mudgal et al., 2010). Without top management commitment, no quality initiative can succeed. Top management must be convinced that registration and certification will enable the organization to demonstrate to its customers a visible commitment to quality. The top management should provide evidence of its commitment to the development and implementation of the quality management system and continually improve its effectiveness. The lack of management support is attributed to management not completely understanding the true goal of the implementation of 5S (Chan et al., 2005; Rodrigues and Hatakeyama, 2006 and Attri et al., 2012). 1651 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 4.2 Financial Constraints Financial constraints are a key barrier to implementation of 5S. Information and technological systems require more funds because without these, implementation of 5S is not possible in the present environment. Funds are needed to institute training programs, provide quality resources, payments for external consultants, payment for auditors, and payment for certification. Lack of financial support affects the certification programs also. 5S improves operational efficiency, customer service, provides an ability to focus on core business objectives and provides greater flexibility (Barve et al., 2007). If any organization has insufficient financial resource, it will not be in their budget to implementation of 5S. 4.3 Lack of Awareness of 5S A major barrier of 5S seen in Indian organisation is lack of awareness about the benefits of the 5S provider. The lack of awareness of the benefits of the 5S both from economic and agile could be a major factor for the resistance to change to 5S. But in manufacturing organisation, due to lack of awareness of 5S, manufacturing organisation is not able to improve performance and work efficiency. If employees of any industry will not have proper understanding of 5S. They will not achieve their objective and goals. Better understanding will help in implementation of 5S. Thus, we can say that lack of awareness of 5S is a major barrier to implementation of 5S. 4.4 Lack of Strategic Planning of 5S Strategic planning is the identification of 5S goals and the specification of the long term plan for managing them. The main role of strategic planning is of paramount importance to any new concept to get institutionalized and incorporated into routine business. 4.5 Lack of Employee Commitment Employee check the deadlines and result of the continuous improvement activities, it spend time helping root cause solving problem activities and standardization. Employee has good communication skills and sufficient knowledge about the implementation of 5S. Employee has confidence on 5S implementation and its result. Without employee commitment, no quality initiative can succeed. 4.6 Resistance to Change and Adoption To attenuate this resistance to change, employees at all levels of the organization must be educated about the goals of 5S implementation well in time (Khurana, 2007). A chief barrier seen in implementation of 5S is the resistance to change, human nature being a fundamental barrier. Employee’s commitment to change programmes is essential given that they actually execute implementation activities (Hansson et al., 2003). 4.7 Lack of Cooperation/Teamwork The success of any type of business relies on the effective teamwork and collaboration of employees at all levels of the organisation. When employees fail to work together as a team, business initiatives and goals become more difficult to attain and the surroundings workplace environment can become negative and disrupting. 4.8 Lack of Education and Training Employees of an organization must be properly educated and trained in a sufficient manner. If employees are not trained, this factor also affects the implementation of 5S. Without proper knowledge they will not be aware 1652 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 of work culture of this quality program and many misconceptions will be in their mind without training. According to Mital et al. (1999), there is a dire need to train workers in manufacturing organizations and thereby improve the overall effectiveness and efficiency of such organizations. A long-term educational and training programme should be designed and implemented to train the employees so that reallocation of the human resource for jobs requiring advanced technical expertise can be made possible (Hung and Chu, 2006). 4.9 Lack of Motivation Motivation derived from the latin word ‘movere’, which means to move (Kretiner, 1998). It is an inner drive and intention that causes a person to do something or get a certain way (Fuhrmann, 2006). Employee motivation is a major factor in the success or failure for any organization (Barrs, 2005). Employees as the bridge to Competitiveness So Organizations must invest in effective strategies to motivate employees (Latt, 2008). Motivation changes the behavior of an employee towards work from negative to positive. Besides that, without motivated employees, profit, products, services, moral and productivity suffer. 4.10 Inability to Change Organisational Culture The organisational culture provides the rule of behavior and attitude. Organisational culture is also motivating the employees and helps leaders accelerate strategy implementation in their organisation. Top management is able to change organisational culture for improving performance and work efficiency. 4.11 Non – Clarity of Organisation Policy and 5S Programme Managers and employee have non - clarity of organisation policy and objectives of 5S. Managers have sufficient technical knowledge and skills to manage their employees and clarify organisation policy for achieving their goals and objective. 4.12 Lack of Communication Employees are not communicating about the continuous improvement result, the activities being under taken, the people that was part of the activities, the objectives and next steps. Communication is very essential in any organization. One department has to communicate to the other to get some information. So the relations between the departments should be good otherwise it will harm the effectiveness of the organization. Lack of communication will also result in the non- participation of the employee. An effective communication channel is required in the organization. 4.13 No Proper Vision and Mission For good implementation of 5S any organization must have a proper vision and mission. In any type of an organization without an aim will not be able to take advantage of the quality program. An organization should be clear in this aspect that why they are implementing 5S and what are their objectives or goals. 4.14 Lack of Leadership Leadership relating to quality is the ability to inspire people to make a total willing and voluntary commitment to accomplishing or exceeding organizational goals. Leadership establishes the unity and purpose for the internal environment of the organization. 5S may fail due to lack of leadership. 1653 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 4.15 Conflict with Other Quality Management System Quality management system is a technique used to communicate to employees. Each employee has different opinion regarding the quality management system so there is a conflict among the employee. The rule and regulation also vary from one quality system to another. V. CONCLUSION The present study conclude that 5S is an important tool to organize the whole organization in a systematic manner. 5S satisfies both organization and customer. Implantation of 5S method is the responsibility of the management and the entire team of employees. In this study we identify the different barriers that hinder the implementation of 5S in an organization. With the help of these barriers will improve performance and efficiency in any type of organization. These barriers are independent to each other so with the help of these barriers we improve quality and performance. REFERENCES [1]. Ahuja, I., & Khamba, J. (2008). Strategies and success factors for overcoming challenges in TPM implementation in Indian manufacturing industry. Journal of Quality in Maintenance Engineering, 14(2), 123-147. [2]. Bain, N. (2010). The consultant guide to successfully implementing 5S. Retrieved August, 3, 2010, from http: www.leanjourney.ca/Preview/Preview- TheConsultantsGuideTo5S.pdf [3]. Attri, R., N. Dev, and V. Sharma. 2013. “Interpretive Structural Modelling (ISM) Approach: An Overview.” Research Journal of Management Sciences 2 (2): 3–8. [4]. Achanga, P., E. Shehab, R. Roy, and G. Nelder. 2006. “Critical Success Factors for Lean Implementation within SMEs.” Journal of Manufacturing Technology Management 17 (4): 460–471. [5]. Ablanedo-Rosas, J., B. Alidaee, J. C. Moreno, and J. Urbina. 2010. “Quality Improvement Supported by the 5S, an Empirical Case Study of Mexican Organisations.” International Journal of Production Research 48 (23): 7063–7087. [6]. Anand, G., & Kodali, R. (2010). Development of a framework for implementation of lean manufacturing systems. International Journal of Management Practice, 4(1), 95–116. [7]. Dahlgaard, J.J., & Dahlgaard-Park, S.M. (2006). Lean production, six sigma quality, TQM and company culture. The TQM Magazine, 18(3), 263–281. [8]. Hospital .(2001). Report on the 5s training in provincial hospitals of three convergence sites. Unpublished report: Negros Oriental Provincial Hospital. [9]. Kumar, S., & Harms, R. (2004). Improving business processes for increased operational efficiency:a case study. Journal of Manufacturing Technology Management, 15(7), 662–674. [10]. Ramlall, S. (2004). A Review of Employee Motivation Theories and their Implications for Employee Retention within Organizations . The Journal of American Academy of Business, 52-63. [11]. Raj, T., Shankar, R. and Suhaib, M. (2009) ‘An ISM approach to analyse interaction between barriers of transition to Flexible Manufacturing System’, Int. J. Manufacturing Technology and Management, Vol. [12]. Raj T. and Attri R., Identification and modelling of barriers in the implementation of TQM, International Journal of Productivity and Quality Management, 28(2), 153-179 (2011). 1654 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 [13]. Attri R., Grover S., Dev N. and Kumar D., An ISM approach for modelling the enablers in the implementation of Total Productive Maintenance (TPM), International Journal System Assurance Engineering and Management, DOI: 10.1007/s13198-012-0088-7 (2012) 16. Attri R., Grover S., Dev N. and Kumar D., An. [14]. Ab Rahman, M.N., et al., Implementation of 5S Practices in the Manufacturing Companies: A Case Study. American Journal of Applied Sciences, 2010. 7(8): p. 1182-1189. [15]. Hough, R. (2008). 5S implementaion methodology. Management Services, 35(5)44-45. Retrieved from Academic Search Complete database. [16]. Gapp, R., R. Fisher, and K. Kobayashi. 2008. “Implementing 5S within a Japanese Context: An Integrated Management System.” Management Decision 46 (4): 565–579. [17]. Hirano, H. 1996. 5S for Operators. 5 Pillars of the Visual Workplace. Tokyo: Productivity Press. [18]. Ramos Alonso, L. O. 2002. “Cultural Impact on Japanese Management. An Approach to the Human Resources Management.” [La incidencia cultural en el management japonés. Una aproximación a la gestión de los recursos humanos]. PhD diss., Universidad de Valladolid, Valladolid, Spain. [19]. Ho, S. K. M. 1998. “5-S Practice: A New Tool for Industrial Management.” Industrial Management & Data Systems 98 (2): 55–62. [20]. Liker, J.K., 2004. The Toyota way: fourteen management principles from the world’s greatest manufacturer. New York: McGraw-Hill. [21]. Liker, J.K. and Hoseus, M., 2008. Toyota culture: the heart and soul of the Toyota way. New York: McGraw-Hill. 1655 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 VARIOUS WAVELET ALGORITHMS FOR IRIS RECOGNITION R.Subha1, Dr. M.Pushparani2 1 Research Scholar, Mother Teresa Women’s University, Kodaikanal 2 Professor and Head, Department of Computer Science, Mother Teresa Women’s University, Kodaikanal (India) ABSTRACT Individual identification has become the need of modern day life. The recognition must be high-speed, automatic and infallible. Biometrics has emerged as a strong alternative to identify a person compared to the traditional ways. Also biometric identification can be made speedy, automatic and is already foolproof. Among other biometrics, Iris recognition has emerged as a strong way of identifying any person. Iris recognition is one of the newer biometric technologies used for personal identification. It is one of the most reliable and widely used biometric techniques available. In geAneral, a typical iris recognition method includes Localization, Normalization, and Matching using traditional and statistical methods. Each method has its own strengths and limitations. In this paper, we compare the recital of various wavelets for Iris recognition algorithm like complex wavelet transform, Gabor wavelet, and discrete wavelet transform. Keywords: Iris Recognition, Complex Wavelets, Gabor Wavelets, Discrete Wavelet Transform I. INTRODUCTION Biometric is the science of recognizing a person based on physical or behavioral characteristics. The commonly used biometric features include speech, fingerprint, face, voice, hand geometry, signature, DNA, Palm, Iris and retinal identification future biometric identification vein pattern identification, Body odor identification, Ear shape identification, Body salinity (salt) identification. Out of all biometric system, Iris biometric is best suitable as it cannot be stolen or cannot be easily morphed by any person. The human iris is an annular part between the pupil and white sclera has an extraordinary structure. Fig 1: The Human Eye Structure The iris begins to form in the third month of gestation and structures creating its pattern are largely complete by the eight months, although pigment accretion can continue in the first postnatal years. Its complex pattern can contain many distinctive features such arching ligaments, furrows, ridges, crypts, rings corona, and freckles. These visible characteristics, which are generally called the texture of the iris, are unique to each subject. 1656 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 II. OVERVIEW OF THE IRIS RECOGNITION SYSTEM Biometrics is automated method of identifying a person or verifying the identity of a person based on physiological or behavioral characteristic. Examples of physiological characteristics include hand or fingers images, facial characteristic and iris recognition. Behavioral Image processing techniques can be employed to extract the unique iris pattern from a digitized image of the eye and encode it into the biometric template, which can be stored in database. The iris being a protected internal organ, whose random texture is most reliable and stable throughout life, can serve as a kind of living password that one need not to remember but always carries along. Every iris is distinct, even two irises of the same individual, and the irises of twins are different. Iris patterns are formed before birth and do not change over the course of a life time. This biometric template contains an objective mathematical representation of the unique information stored in the iris, and allows the comparisons made between templates. When a person wishes to be identified by an iris recognition system, their eye is first photographed and then template is created for their iris region. This template is then compared with the template stored in a database, until either a matching template is found and a subject is identified, or no match is found and subject remains unidentified. III. VARIOUS WAVELETS TRANSFORM 3.1 Complex Wavelets Complex Wavelets Transforms use complex valued filtering that decomposes the real/complex signals into real and imaginary parts in transform domain. The real and imaginary coefficients are used to compute amplitude and phase information, just the type of information needed to accurately describe the energy localization of oscillating functions. Here complex frequency B-spline wavelet is used for iris feature extraction. 3.1.1 Result The iris templates are matched using different angles 210,240,280,320 and 350 Degrees and it is observed that as angles increases percentage of matching also increases the better match is observed at angle 350 which is 93.05%.Further by detecting eyelids and eyelashes the iris image is cropped and iris template is generated for matching purpose the results obtained is better than previous results the matching score is 95.30%. 3.2 Gabor Wavelet The main idea of this method is that: firstly we construct two-dimensional Gabor filter, and we take it to filter these images, and after we get phase information, code it into 2048 bits, i.e. 256 bytes. In image processing, a Gabor filter, named after Dennis Gabor, is a linear filter used for edge detection. Frequency and orientation representations of Gabor filter are similar to those of human visual system, and it has been found to be particularly appropriate for texture representation and discrimination. In the spatial domain, a 2D Gabor filter is a Gaussian kernel function modulated by a sinusoidal plane wave. The Gabor filters are selfsimilar - all filters can be generated from one mother wavelet by dilation and rotation. Its impulse response is defined by a harmonic function multiplied by a Gaussian function. Because of the multiplication-convolution property (Convolution theorem), the Fourier transform of a Gabor filter's impulse response is the convolution of the Fourier transform of the harmonic function and the Fourier 1657 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 transform of the Gaussian function. The filter has a real and an imaginary component representing orthogonal directions. The two components may be formed into a complex number or used individually. Gabor filters are directly related to Gabor wavelets, since they can be designed for a number of dilations and rotations. However, in general, expansion is not applied for Gabor wavelets, since this requires computation of bi-orthogonal wavelets, which may be very time- consuming. Therefore, usually, a filter bank consisting of Gabor filters with various scales and rotations is created. The filters are convolved with the signal, resulting in a so-called Gabor space. This process is closely related to processes in the primary visual cortex. Jones and Palmer showed that the real part of the complex Gabor function is a good fit to the receptive field weight functions found in simple cells in a cat's striate cortex. 3.2.1 Result We use the Daugman's methods to iris regions segmentation and use Gabor wavelet for feature extraction. At last, in the identification stage we calculate Hamming distance between a test image & a training image. The smallest distance among them is expressed, that test image belongs to this class. The recognition rate is 96.5%. 3.3 The Discrete Wavelet Transform Computing wavelet coefficients at every possible scale is a fair amount of work, and it generates an awful lot of data. That is why we choose only a subset of scales and positions at which to make our calculations. It turns out, rather remarkably, that if we choose scales and positions based on powers of two so-called dyadic scales and positions then our analysis will be much more efficient and just as accurate. 3.3.1 Result The technique developed here uses all the frequency resolution planes of Discrete Wavelet Transform (DWT). These frequency planes provide abundant texture information present in an iris at different resolutions. The accuracy is improved up to 98.98%. With proposed method FAR and FRR is reduced up to 0.0071% and 1.0439% respectively. IV. CONCLUSION In this paper, we compare the performance of various wavelets for Iris recognition like complex wavelet transform, Gabor wavelet, and discrete wavelet transform. Using complex wavelet, different coefficient vectors are calculated. Minimum distance classifier was used for final matching. The smaller the distance the more the images matched. It is observed that for the complex wavelets the results obtain are good than the simple wavelet because in complex wavelet we get both phase and angle also real and imaginary coefficients, so we can compare all these parameters for iris matching purpose.2D Gabor wavelets have the highest recognition rate. Because iris is rotator, and 2D Gabor wavelets have rotation invariance, it has the highest recognition rate. But 2D Gabor wavelets have high computational complexity, and need more time. Discrete wavelet transform used for iris signature formation gives better and reliable results. REFERENCE [1]. Biometrics: Personal Identification in a Networked Society, A. Jain, R. Bolle and S. Pankanti, eds. Kluwer, 1999. 1658 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 [2]. ISBN: 978-81-931039-3-7 D. Zhang, Automated Biometrics: Technologies and Systems. Kluwer, 2000 Anil Jain. Introduction to Biometrics. Michigan State University, East Lansing, MI. [3]. L. Ma, T, Yunhong Wang, and D. Zhang. Personal identification based on iris texture analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence, vol.25, no.12, 2003 [4]. John Daugman. Recognizing persons by their iris patterns. Cambridge University, Cambridge, UK. [5]. J. Daugman, "Demodulation by Complex-Valued Wavelets for Stochastic Pattern Recognition," Int'l J. Wavelets, Multiresolution and Information Processing, vol. 1, no. 1, pp. 1-17, 2003 [6]. W. Boles and B. Boashash, "A Human Identification Technique Using Images of the Iris and Wavelet Transform," IEEE Trans. Signal Processing, vol. 46, no. 4, pp. 1185- 1188, 1998 [7]. R. Wildes, J. Asmuth, G. Green, S. Hsu, R. Kolczynski, J. Matey, and S. McBride, "A Machine-Vision System for Iris Recognition," Machine Vision and Applications, vol. 9, pp. 1-8, 1996 [8]. Makram Nabti and Bouridane, "An effective iris recognition system based on wavelet maxima and Gabor filter bank", IEEE trans. on iris recognition, 2007. [9]. Narote et al. "An iris recognition based on dual tree complex wavelet transform". IEEE trans. on iris recognition, 2007. [10]. Institute of Automation Chinese Academy of Sciences. Database of CASIA iris image [EB/OL] [11]. L. Masek, "Recognition of Human Iris Patterns for Biometric Identification", The University of Western California, 2003. [12]. N. G. Kingsbury, "Image processing with complex wavelets," Philos.Trans. R. Soc. London A, Math. Phys. Sci, vol. 357, no. 3, pp. 2543-2560, 1999. [13]. Vijay M.Mane, GauravV. Chalkikar and Milind E. Rane, "Multiscale Iris Recognition System", International journal of Electronics and Communication Engineering & Technology (IJECET), Volume 3, Issue 1, 2012, pp. 317 - 324, ISSN Print: 0976- 6464, ISSN Online: 0976 -6472. [14]. Darshana Mistry and Asim Banerjee, "Discrete Wavelet Transform using Matlab", International journal of Computer Engineering & Technology (IJCET), Volume 4, Issue 2, 2012, pp. 252 - 259, ISSN Print: 0976 - 6367, ISSN Online: 0976 - 6375. [15]. Sayeesh and Dr. Nagaratna p. Hegde “A comparison of multiple wavelet algorithms for iris Recognition” International Journal of Computer Engineering and Technology (IJCET), ISSN 09766367(Print), ISSN 0976 - 6375(Online) Volume 4, Issue 2, March - April (2013), © IAE 1659 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 HIDING THE DATA USING STEGANOGRAPHY WITH DIGITAL WATERMARKING 1 G.Thirumani Aatthi, 2A.Komathi 1 Research Scholar, Department of Computer Science & Information Technology, Nadar Saraswathi College of Arts & Science, Theni,Tamil Nadu,(India) 2 Department of Computer Science & Information Technology; Nadar Saraswathi College of Arts & Science, Theni, Tamil Nadu,(India) ABSTRACT Data Security is the method of shielding Information. It protects its accessibility, privacy and Integrity. Access to Stored data on computer data base has improved greatly. More Companies store business and individual information on computer than ever before. Much of the data stored is highly confidential and not for public viewing. Cryptography and steganography are well known and widely used techniques that manipulate information in order to cipher or hide their existence. These two techniques share the common goals and services of protecting the confidentiality, integrity and availability of information from unauthorized access. In Existing research, data hiding system that is based on image steganography and cryptography is proposed to secure data transfer between the source and destination. In this the main drawback was that, the hackers may also get the opportunity to send some information to destination and it may lead confusion to receiver. In my research I proposed LSB(Leased Significant Bit) Technique used for finding the image pixel position and pseudorandom permutation method used for store the data in random order. Moreover I have proposed digital watermark technique to avoid the unauthorized receiving information from hackers. In this proposed system, the above three technique will be combined for secure data transfer. Experimental results will prove the efficiently and security of my Proposed work. Key Word: Cryptography, Steganography, Data Security, Key Generation I. INTRODUCTION In the Internet one of the most important factors of information technology and communication has been the security of information. Cryptography was created as a technique for securing the secrecy of communication and many different methods have been developed to encrypt and decrypt data in order to keep the message secret. Unfortunately it is sometimes not enough to keep the contents of a message secret, it may also be necessary to keep the existence of the message secret. The technique used to implement this, is called steganography. Steganography is the art and science of invisible communication. This is accomplished through hiding information in other information, thus hiding the existence of the communicated information. All digital file formats can be used for steganography, the four main categories of file formats that can be used for steganography. 1660 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 Figure 1: Categories of Steganography Over the past few years, numerous steganography techniques that embed hidden messages in multimedia objects have been proposed. There have been many techniques for hiding information or messages in images in such a manner that the alterations made to the image are perceptually indiscernible. Common approaches are include: (i) Least significant bit insertion (LSB) (ii) Masking and filtering (iii) Transform techniques Least significant bits (LSB) insertion is a simple approach to embedding information in image file. The simplest steganographic techniques embed the bits of the message directly into least significant bit plane of the coverimage in a deterministic sequence. Modulating the least significant bit does not result in human-perceptible difference because the amplitude of the change is small. Masking and filtering techniques performs analysis of the image, thus embed the information in significant areas so that the hidden message is more integral to the cover image than just hiding it in the noise level. Transform techniques embed the message by modulating coefficients in a transform domain, such as the Discrete Cosine Transform (DCT) used in JPEG compression, Discrete Fourier Transform, or Wavelet Transform. These methods hide messages in significant areas of the cover-image, which make them more robust to attack. Transformations can be applied over the entire image, to block through out the image, or other variants. II. LITERATURE REVIEW Here, Ms.Dipti and Ms.Neha, developed a technique named, “Hiding Using Cryptography and Steganography”[01]is discussed. In that, Steganography and Cryptography are two popular ways of sending vital information in a secret way. One hides the existence of the message and the other distorts the message itself. There are many cryptography techniques available; among them AES is one of the most powerful techniques. In Steganography we have various techniques in different domains like spatial domain, frequency domain etc. to hide the message. It is very difficult to detect hidden message in frequency domain and for this domain we use various transformations like DCT, FFT and Wavelets etc. They are developing a system where we develop a new technique in which Cryptography and Steganography are used as integrated part along with newly developed enhanced security module. In Cryptography we are using AES algorithm to encrypt a message and a part of the message is hidden in DCT of an image; remaining part of the message is used to generate two secret keys which make this system highly secured. Another newly developed Method named,“Two New Approaches for Secured Image Steganography Using Cryptographic Techniques and Type Conversions. Signal & Image Processing” [02] is discussed. Mr. Sujay, N. and Gaurav, P,introduces two new methods wherein cryptography and steganography are combined to encrypt the data as well as to hide the encrypted data in another medium so the fact that a message being sent is concealed. One of the methods shows how to secure the image by converting it into cipher text by S-DES 1661 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 algorithm using a secret key and conceal this text in another image by steganographic method. Another method shows a new way of hiding an image in another image by encrypting the image directly by S-DES algorithm using a key image and the data obtained is concealed in another image. Another newly developed Method named, “ Image Based Steganography and Cryptography “[06] is discussed. In that Mr. Domenico, B. and Luca, L. year.. describe a method for integrating together cryptography and steganography through image processing. In particular, they present a system able to perform steganography and cryptography at the same time using images as cover objects for steganography and as keys for cryptography. They will show such system is an effective steganographic one (making a comparison with the well known F5 algorithm) and is also a theoretically unbreakable cryptographic one (demonstrating its equivalence to the Vernam Cipher) III. PROPOSED WORK In this paper I proposed, Data hiding in media, including images, video, and audio, as well as in data files is currently of great interest both commercially, mainly for the protection of copyrighted digital media, and to the government and law enforcement in the context of information systems security and covert communications. So I present a technique for inserting and recovering “hidden” data in image files as well as gif files. Each Color pixel is a combination of RGB Values wherein each RGB components consists of 8 bits. If the letters in ASCII are to be represented within the color pixels, the rightmost digit, called the Least Significant Bit (LSB) can be altered to hide the images. 3.1 Key Stream Generation In order to encrypt the message, we choose a randomly generated key-stream. Then the encryption is done byte by byte to get the ciphered text. The key stream is generated at the encryption and decryption site. For encryption, a secret seed is applied to the content which in turn generates the key stream. In order to generate the same key at the decryption site, the seed must be delivered to the decryption site through a secret channel.Once the seed is received, it can be applied to the cipher text to generate the key stream which is further used for decryption C = E(M,K) = (Mi+Ki) mod 255 where i=0 to L-1 Where M is Message K is randomly generated key-stream. 3.2 Watermarking Algorithm 3.2.1 SS (Spread Spectrum) We proposed a spread spectrum watermarking scheme. The embedding process is carried out by first generating the watermark signal by using watermark information bits, chip rate and PN sequence.The watermark information bits b = {bi}, where bi = {l, -1}, are spread by, which gives aj=bi The watermark signal W = {wj}, where wj = ajPj where Pj= {l, -1} The watermark signal generated is added to the encrypted signal, to give the watermarked signal CW= C + W = Cwi= ci + wi. 1662 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 IV. SAMPLE SCREEN SHOTS Send the Message Receive the Message Encrypt & Validation Code Verification Encryption Key & Validation Code Decrypt: (Water Marking With Text) V. CONCLUSION I propose a novel technique to embed a robust watermark in the JPEG2000 compressed encrypted images using three different existing watermarking schemes. The algorithm is simple to implement as it is directly performed in the compressed-encrypted domain, i.e., it does not require decrypting or partial decompression of the content. 1663 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 The scheme also preserves the confidentiality of content as the embedding is done on encrypted data. The homomorphic property of the cryptosystem are exploited, which allows us to detect the watermark after decryption and control the image quality as well. The detection is carried out in compressed or decompressed domain. In case of decompressed domain, the non-blind detection is used. I analyze the relation between payload capacity and quality of the image (in terms of PSNR and SSIM) for different resolutions. Experimental results show that the higher resolutions carry higher payload capacity without affecting the quality much, whereas the middle resolutions carry lesser capacity and the degradation in quality is more than caused by watermarking higher resolutions. REFERENCES [1]. Dipti, K. S. and Neha, B. 2010. Proposed System for Data Hiding Using Cryptography and Steganography. International Journal of Computer Applications. 8(9), pp. 7-10. Retrieved 14th August, 2012 [2]. Sujay, N. and Gaurav, P. 2010. Two New Approaches for Secured Image Steganography Using Cryptographic Techniques and Type Conversions. Signal & Image Processing: An International Journal (SIPIJ), 1(2), pp 60-73. [3]. Domenico, B. and Luca, L. year. Image Based Steganography and Cryptography. [4]. Jonathan Cummins, Patrick Diskin, Samuel and Robert Par-lett,“Steganography and Digital Watermarking”, 2004. [5]. Clara Cruz Ramos, Rogelio Reyes Reyes, Mariko Nakano Miyata-keandHéctor Manuel Pérez Meana, “Watermarking-Based Image Authentication System in the Discrete Wavelet Transform Domain”.intechopen. [6]. Domenico, B. and Luca, L. year. Image Based Steganography and Cryptography. [7]. Niels, P. and Peter, H 2003. Hide and Seek: An Introduction to Steganography. IEEE Computer Society. IEEE Security and Privacy, pp. 32-44. 1664 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 IMPROVING DATA SECURITY IN AD-HOC NETWORK BASED ON CPHS ALGORITHM N.Vandhana1, P.Nithya2 1 Research Scholar, Department of Computer Science & Information Technology Nadar Saraswathi College of Arts & Science, Theni,Tamil Nadu, (India) 2 Department of Computer Science & Information Technology Nadar Saraswathi College of Arts & Science, Theni,Tamil Nadu, (India) ABSTRACT An ad-hoc network is a temporary network connection created for a specific purpose such as transferring data from one computer to another. Ad-hoc networks confirmed their efficiency being used in different fields but they are highly vulnerable to security attacks and dealing with this is one of the main challenges of these networks today. In Existing System, the data transmission is based only on the encryption technique and follows the same arrangement of the data frame. In this the main drawback was that the hackers may also know the arrangement of the data frame. So the data will be easily hacked or damaged by the hacker. In this research, the proposed CPHS (Crypto Puzzle Hiding System) algorithm is based on key generation algorithm, used for encrypt the data and change the arrangement of the data frame. And the hackers will be confused to decrypt the data. So the research will provide the security for data transmission. Experiments results will prove the efficient and security of my proposed work. I. INTRODUCTION The open nature of the wireless medium leaves it vulnerable to intentional interference attacks, typically referred to as jamming. This intentional interference with wireless transmissions can be used as a launch pad for mounting Denial-of-Service attacks on wireless networks. Typically, jamming has been addressed under an external threat model. However, adversaries with internal knowledge of protocol specifications and network secrets can launch low-effort jamming attacks that are difficult to detect and counter. In this work, we address the problem of selective jamming attacks in wireless networks. In these attacks, the adversary is active only for a short period of time, selectively targeting messages of high importance. We illustrate the advantages of selective jamming in terms of network performance degradation and adversary effort by presenting two case studies; a selective attack on TCP and one on routing. We show those selective jamming attacks can be launched by performing real-time packet classification at the physical layer. To mitigate these attacks, we develop three schemes that prevent real-time packet classification by combining cryptographic primitives with physical layer attributes. We analyze the security of our methods and evaluate their computational and communication overhead. Wireless networks rely on the uninterrupted availability of the wireless medium to interconnect participating nodes. However, the open nature of this medium leaves it vulnerable to multiple security threats. Anyone with a transceiver can eavesdrop on wireless transmissions, inject spurious messages, or jam legitimate ones. While eavesdropping and message injection can be prevented using cryptographic methods, jamming attacks are much 1665 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 harder to counter. They have been shown to actualize severe Denial-of-Service (DoS) attacks against wireless networks. In the simplest form of jamming, the adversary interferes with the reception of messages by transmitting a continuous jamming signal, or several short jamming pulses. Typically, jamming attacks have been considered under an external threat model, in which the jammer is not part of the network. Under this model, jamming strategies include the continuous or random transmission of high-power interference signals. However, adopting an “always-on” strategy has several disadvantages. First, the adversary has to expend a significant amount of energy to jam frequency bands of interest. Second, the continuous presence of unusually high interference levels makes this type of attacks easy to detect. Conventional anti-jamming techniques rely extensively on spread-spectrum (SS) communications, or some form of jamming evasion (e.g., slow frequency hopping, or spatial retreats). SS techniques provide bit-level protection by spreading bits according to a secret pseudo-noise (PN) code, known only to the communicating parties. These methods can only protect wireless transmissions under the external threat model. Potential disclosure of secrets due to node compromise neutralizes the gains of SS. Broadcast communications are particularly vulnerable under an internal threat model because all intended receivers must be aware of the secrets used to protect transmissions. Hence, the compromise of a single receiver is sufficient to reveal relevant cryptographic information. In this paper, we address the problem of jamming under an internal threat model. II. RESULT AND DISCUSSION 2.1. Hiding Based On Cryptographic Puzzles We present a packet hiding scheme based oncryptographic puzzles. The main idea behind such puzzlesis to force the recipient of a puzzle execute a pre-definedset of computations before he is able to extract a secret ofinterest. The time required for obtaining the solution ofa puzzle depends on its hardness and the computational ability of the solver. The advantage of the puzzlebased scheme is that its security does not rely on the PHYlayer parameters. However, it has higher computation andcommunication overhead. Fig 2: The Cryptographic Puzzle Based Hiding Scheme In our context, we use cryptographic puzzles to temporary hide transmitted packets. A packet m is encryptedwith a randomly selected symmetric key k of a desirablelength s. The key k is blinded using a cryptographic puzzleand sent to the receiver. For a computationally boundedadversary, the puzzle carrying k cannot be solved beforethe transmission of the encrypted version of m is completedand the puzzle is received. Hence, the adversary cannotclassify m for the purpose of selective jamming. 1666 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 2.1.1. Cryptographic Puzzle Hiding Scheme (CPHS) Let a sender S have a packet m for transmission. The sender selects a random key k ∈ {0, 1} s , of a desired length. S generates a puzzle P = puzzle(k, tp), where puzzle() denotes the puzzle generator function, and tp denotes the time required for the solution of the puzzle. Parameter tp is measured in units of time, and it is directly dependent on the assumed computational capability of the adversary, denoted by N and measured in computational operations per second. After generating the puzzle P , the sender broadcasts (C, P ), where C = Ek (π 1(m)). At the receiver side, any receiver R solves the received puzzle P to recover key k ′ and then computes m′= π −1 (Dk′ (C′)). If the decrypted packet m′ is meaningful (i.e., is in the proper format, has a valid CRC code, and is within the context of the receiver’s communication), the receiver accepts that m′ = m. Else, the receiver discards m′. Fig. 2 shows the details of CPHS. III. CONCLUSION AND FUTURE ENHANCEMENT CONCLUSION We addressed the problem of selective jamming attacks in wireless networks. We considered an internal adversary model in which the jammer is part of the network under attack, thus being aware of the protocol specifications and shared network secrets. We showed that the jammer can classify transmitted packets in real time by decoding the first few symbols of an ongoing transmission. We evaluated the impact of selective jamming attacks on network protocols such as TCP and routing. Our findings show that a selective jammer can significantly impact performance with very low effort. We developed three schemes that transform a selective jammer to a random one by preventing real-time packet classification. Our schemes combine cryptographic primitives such as commitment schemes, cryptographic puzzles, and all-or-nothing transformations (AONTs) with physical layer characteristics. We analyzed the security of our schemes and quantified their computational and communication overhead. BIBLIOGRAPHY [1]. T. X. Brown, J. E. James, and A. Sethi. Jamming and sensing of encrypted wireless ad hoc networks. In Proceedings of MobiHoc, pages 120–130, 2006. [2]. M. Cagalj, S. Capkun, and J.-P. Hubaux. Wormhole-based antijamming techniques in sensor networks. IEEE Transactions on MobileComputing, 6(1):100–114, 2007. [3]. A. Chan, X. Liu, G. Noubir, and B. Thapa. Control channel jamming: Resilience and identification of traitors. In Proceedings of ISIT, 2007. [4]. T. Dempsey, G. Sahin, Y. Morton, and C. Hopper. Intelligent sensing and classification in ad hoc networks: a case study. Aerospace andElectronic Systems Magazine, IEEE, 24(8):23–30, August 2009. [5]. Y. Desmedt. Broadcast anti-jamming systems. Computer Networks, 35(2-3):223–236, February 2001. [6]. K. Gaj and P. Chodowiec. FPGA and ASIC implementations of AES. Cryptographic Engineering, pages 235–294, 2009. [7]. O. Goldreich. Foundations of cryptography: Basic applications. Cambridge University Press, 2004. [8]. B. Greenstein, D. Mccoy, J. Pang, T. Kohno, S. Seshan, and D. Wetherall. Improving wireless privacy with an identifier-free link layer protocol. In Proceedings of MobiSys, 2008. [9]. A. Juels and J. Brainard. Client puzzles: A cryptographic countermeasure against connection depletion attacks. In Proceedings of NDSS, pages 151–165, 1999. 1667 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 [10]. Y. W. Law, M. Palaniswami, L. V. Hoesel, J. Doumen, P. Hartel, and P. Havinga. Energy-efficient linklayer jamming attacks against WSN AC protocols. ACM Transactions on Sensors Networks, 5(1):1–38, 2009. WEB SITES 1. http://www.sourcefordgde.com 2. http://www.networkcomputing.com/ 3. http://www.almaden.ibm.com/software/quest/Resources/ 4. http://oceanstore.cs.berkely.edu 5. http://ww.amazon.com/gp/browse amazing store 6. http://en.amazingstore.org/,2013 7. clever safe inc., http://www.cleversafe.org/dispersed-storage,2013 8. http://www.overnet.com 9. http://www.netcraft.com 10. http://ramp.uecd.edu/projects/recall 11. http://www.seagate.com/support/kd/disc/smart.html,2013 12. http://simpy.sourceforge.ref/(2013) 1668 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 DESIGN AND IMPLEMENTATION OF ADAPTIVE ECHO CANCELLER BASED LMS & NLMS ALGORITHM Sandip A. Zade1, Prof. Sameena Zafar2 1 Mtech student,Department of EC Engg., Patel college of Science and Technology Bhopal(India) 2 Acadamic Dean. Patel College of Science and Technology Bhopal(India) ABSTRACT This paper proposes the adaptive echo cancellation using normalized least mean square (NLMS) algorithm. The NLMS algorithm is the normalized version of least mean square (LMS) algorithm. Todays technical scenario communication system, possesses additive noise, signal interference and echo etc. Due to this reason the error is generated at the time of transmission of data. Hence adaptive filter is a appropriate option to reduce the noise or channel effects. In this paper high speed of adder, substractor and Adaptive filter coefficients to design LMS and NLMS algorithm is realized. This paper focuses on the use of LMS and NLMS algorithms to reduce this unwanted echo, thus increasing communication quality. This work better highlights differences between algorithm performance than previously published work and sheds new light on algorithm behavior in terms of SNR. This work were created in the popular VHDL language. The NLMS algorithm establishes a better balance between simplicity and performance than least mean square algorithm. Keywords: Adaptive Filter, NLMS Algorithm, LMS Algorithm, Vhdl Language I. INTRODUCTION An echo is a reflection of sound, arriving at the listener some time after the direct sound. Echo is the reflected copy of the voice heard some time later and delayed version of the original. The term echo cancellation is used in telephony to describe the process of removing echo from a voice communication in order to improve voice quality on a telephone call. Echo cancellation is the process which removes unwanted echoes from the signal on a telephone line. It includes first recognizing the originally transmitted signal that re-appears, with some delay, in the transmitted or received signal. Once the echo is recognized, it can be removed by 'subtracting' it from the transmitted or received signal. Multiple reflections in acoustic enclosures and transmission delay affect the sound quality, which in the case of a teleconferencing system lead to a poor understanding of the conversation. In addition to improving subjective quality, this process increases the capacity achieved through silence suppression by preventing echo from traveling across a network. Hands-free phone is a basic and intrinsic application with small information terminals such as cell phones, smart phones, and tablet PCs. For hands-free communication, echo cancellation is common but still a difficult function. An echo canceller has an adaptive filter to emulate the echo path between the input of amplifier to drive a loudspeaker and the microphone. Even with echo cancellers, suppression of echo is very difficult because the loudspeaker is small and close to the 1669 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 microphone, and the sound from the loudspeaker is very loud. Echo cancellation try to eliminate the echo from the transmitted audio signal with a special optimization algorithm. The algorithm produces copies of the received signal and checks for parts of the signal that reappear with some delay. This reappearing parts are then subtracted from the signal, the echo is removed. In this paper, the LMS and NLMS adaptive algorithms are described. The following sub-sections will present a brief description of adaptive filters focused on the cancellation of noise and echo. The second section if focused on briefly review of the adaptive filtering algorithm. The third section presents acoustic echo cancellation. The fourth section presents literature survey. The fifth section shows the methodology used. The last section presents the simulation and experimental results obtained from the adaptive cancellation of noise & it summarizes the main findings of the paper. II. VARIOUS ADAPTIVE FILTERING ALGORITHMS A filter which adapts itself to the input signal given to it. Adaptive Filtering dealing with adaptive filters and system design. The main function of adaptive filtering is the development of a filter capable of adjusting to the statistics of the signal. Usually, an adaptive filter takes the form of a FIR filter, with an adaptive algorithm that modifies the values of its coefficients. They are used in a wide range of applications including system identification, noise cancellation, signal prediction, echo cancellation and adaptive channel equalization, process control. The main configurations of adaptive filters are the adaptive cancellation of noise and echo. For this justification, the filter uses an adaptive algorithm to change the value of the filter coefficients, so that it acquires a better approximation of the signal after each iteration. The LMS, and its variant the NLMS are two of the adaptive algorithms widely in use. Also the normalized version of the LMS algorithm, called Normalized Least Mean Square (NLMS) algorithm is widely used. NLMS algorithm has been used more often in real time applications. The LMS algorithm have slow convergence and poor tracking as compare to the the NLMS algorithm. Both algorithms require a small number of multiplications and additions for the update of the coefficients. 2.1 LMS Algorithm [9] The LMS algorithm adapts the filter tap weights so that e(n) is minimized in the mean-square sense. LMS algorithm is sensitive to variation in step size parameter. LMS algorithm requires number of itrations equal to dimensionality of the input. When the processes x(n) & d(n) are jointly stationary, this algorithm converges to a set of tap-weights which, on average, are equal to the Wiener-Hoff solution. It acts as a negative feedback to minimize error signal. 1. The output of the FIR filter, y(n) is 2. The value of the error estimation is e(n) = d(n) - y(n) 3. The tap weights of the FIR vector are updated in preparation for the next iteration 1670 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 The main reason for the LMS algorithms popularity in adaptive filtering is its computational simplicity, making it easier to implement than all other commonly used adaptive algorithms. The only disadvantage is its weak convergence. For each iteration the LMS algorithm requires 2N additions and 2N+1 multiplication. Where n= 0,1,2,....n. 1.2 NLMS Algorithm [9] As the NLMS is an extension of the standard LMS algorithm, the NLMS algorithms practical implementation is very similar to that of the LMS algorithm. One of the drawback of LMS algorithm is selection of step size parameter. In order to solve this difficulty we can use the NLMS algorithm. Here the step size parameter is normalized.So the NLMS algorithm is time varying step size algorithm. The NLMS algorithm is used in system in order to improve voice quality. Each iteration of the NLMS algorithm requires these steps in the following order 1. The output of the adaptive filter is calculated 2. An error signal is calculated as the difference between the desired signal and the filter output e(n) = d(n) – y(n) 3. The step size value for the input vector is calculated 4. The filter tap weights are updated in preparation for the next iteration w(n+1) = w(n) + µ(n)e(n)x(n) Each iteration of the NLMS algorithm requires 3N+1 multiplications, this is only N more than the standard LMS algorithm. This is an acceptable increase considering the gains in stability and echo attenuation achieve[10]. Where n= 0,1,2,....n. Adaptive filters are systems with four terminals as shown in Fig.1, where x is the input signal, d is the desired signal, y is output signal of filter and e is the error signal [4]. Adaptive filters design technique may be digital, analog or mixed. Every technique presents advantages and disadvantages, for example, analog adaptive filters are very fast, but offset avoids getting the least error [4]. Digital filters are slow but precise, because is necessary the use of a lot of components, due to floating point operations. Mixed design (analog and digital), offers a good compromise between precision and speed, but VLSI design is more complicated because is necessary to separate analog and digital components inside the chip .The LMS algorithm is one of the most used algorithms because it is easy and stable. The only disadvantage is its weak convergence [3]. The length of the FIR filter is 64 and the size of the step is 0.001. Two inputs are required: · A reference noise that should be related with the noise that exists in distorted the input signal. This means that the noise comes from the same source. · An error signal already calculated. 1671 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 Figure 1 Adaptive Filter III. ACOUSTIC ECHO CANCELLATION (AEC) The AECs is a system identification application as shown in Fig.2, which uses adaptive filters to obtain a copy of the acoustic transfer function (the response of an enclosure to an acoustic impulse). Acoustic echo occur when an audio signal is reverberated in a real environment, resulting in the original intended signal plus attenuated and delay time. The signal applied to the loudspeaker(s) x(n) propagates through multiple acoustic paths and it is picked up by the microphone(s). This signal is used as the desired signal d(n) in the system identification process. The output of the adaptive filter y(n) is determined by convolving the samples x(n) with the adaptive filter coefficients w(n).The filter is altered iteratively to minimize the error signale(n). The coefficient update can be carried out with various adaptive algorithms. Figure 2 Acoustic Echo Cancellation Public addressing systems are affected by acoustic feedback that may lead to the saturation of the system. In order to improve the sound quality and prevent audio feedback the acoustic echo cancellers (AECs) are deployed to remove the undesired echoes resulting from the acoustic coupling between the loudspeaker(s) and the microphone(s). IV. LITERATURE SURVEY The paper [2] presents a solution to noise or echo cancellation and a hardware real-time implementation of the LMS algorithm. The overall performance of the designed adaptive filter compared with other implemented systems using the same filter is good, and with further improvements the results will improve. The LMS 1672 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 algorithm provides good numerical stability and its hardware requirements are low. On the other hand, the NLMS algorithm is one of the most implemented adaptive algorithms in actual industrial applications. This paper [3] proposes an FPGA implementation of an Adaptive Noise Canceller using the LMS algorithm. In this paper, in order to show the performance of FPGA in digital signal processing applications, implement an Adaptive Noise Canceller on an FPGA and use the LMS algorithm as the adaptive filtering algorithm of the Adaptive Noise Canceller. The performance of the LMS algorithm implemented by hardware is comprehensively analyzed in terms of convergence performance, truncation effect and tracking ability. In paper [4] an echo canceller is presented, using an adaptive filter with a modified LMS algorithm, where this modification is achieved coding error on conventional LMS algorithm. V. METHODOLOGY USED The architecture of adaptive filter is already implemented with LMS Algorithm for Adaptive echo cancellation. Also the comparison between LMS & NLMS algorithm is being done. The NLMS algorithm outperforms the LMS algorithm, in terms of Mean Square Error. It has a better convergence than LMS algorithm. The NLMS adaptive filtering algorithm is expressed by its simplicity in implementation and its stability. These advantages recommend the NLMS algorithm as a good choice for real time implementation. Hence the NLMS algorithm can be implemented for Adaptive echo cancellation in VHDL. Also the comparison of LMS and NLMS algorithm for echo cancellation along with signal to noise ratio has been done. Fig.3 shows the architecture used. The system architecture consist of main four modules such as 16 bit adder, 16 bit substractor, multiplier and delay. Figure 3 System Architecture Where, x(n),x(n-1),……,x(n-M+1) is the tap inputs x(n) is the reference signal M-1 is the number of delay elements. d(n) denotes the primary input signal e(n) denotes the error signal and wi(n) denotes the tap weight at the nth iteration.. VI. RESULT AND CONCLUSIONS 1673 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 6.1 VHDL Result for Signal 1 Figure 4 Training Signal 1 of LMS algorithm Figure 5 Testing signal 1 of LMS algorithm Figure 6 Training signal 1 of NLMS algorithm Figure 7 Testing signal 1 of NLMS algorithm 1674 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 6.2 MATLAB Results for LMS & NLMS Algorithm Figure 8 Signal 1 of LMS Algorithm Figure 9 Signal 1 of NLMS Algorithm The PSNR of the LMS algorithm is 4.08 dB and the PSNR of the NLMS algorithm is 29.72 dB. Whereas the mean square error of the LMS algorithm is 0.39 and the mean square error of the NLMS algorithm is 0.00106. Figure 10 Signal 2 of LMS algorithm 1675 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 Figure 11 Signal 2 of NLMS Algorithm The PSNR of the LMS algorithm is 2.88 dB and the PSNR of the NLMS algorithm is 22.17 dB. Whereas the mean square error of the LMS algorithm is 0.514 and the mean square error of the NLMS algorithm is 0.006. The adaptive echo canceller based LMS and NLMS algorithm is successfully designed wherein comparison of LMS and NLMS algorithm has been done in terms of PSNR. The PSNR of the NLMS algorithm is 25.945 dB and the PSNR of the LMS algorithm is 3.48dB. So the PSNR is maximum it realize better sound quality. REFERENCES [1] I. Homana, et al, “FPGA Implementation of LMS and NLMS Adaptive Filter for Acoustic Echo Cancellation”, in Journal of Electronics & Telecommunication, vol 52, no. 4, 2011. [2] C. G. Saracin, et al, “Echo Cancellation using LMS Algorithm”, in UPB Science Bulletin Series C,vol 71, no 4, 2009. [3] TianLan, et al, “FPGA Implementation of Adaptive Noise Canceller”, in IEEE International Symposium on Information Processing, 2008. [4] J. V. Lopez, et al, “Adaptive Echo Canceller using Modified LMS Algorithm”, in IEEE International Conference on Electrical & Electronics Engineering, 2005. [5] R. Dony, et. al., “An FPGA Implementation of the LMS Adaptive Filter for Audio Processing”, Proc. of the 12th Annual IEEE Symposium on Field-Programmable Custom Computing Machines (FCCM04), pp. 324- 335, 2006. [6] Elhossini, S. Areibi and R. Dony, “An FPGA Implementation of the LMS Adaptive Filter for Audio Processing,” in Proc. IEEE International Conference on Reconfigurable Computing and FPGAs, Sept. 2006, pp [7] L. K. Ting, R. F. Woods and C. F. N. Cowan, “Virtex FPGA Implementation of a Pipelined Adaptive LMS Predictor for Electronic Support Measures Receivers,” IEEE Trans. VLSI Syst., vol. 13, Jan. 2005, pp. 8695. [8] P. Waldeck and N. Bergmann, “Evaluating software and hardware implementations of signal-processing tasks in an FPGA,” in Proc. IEEE International Conference on Field-Programmable Technology, Brisbane, Australia, Dec. 2004, pp. 299-302. 1676 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 [9] ISBN: 978-81-931039-3-7 Alok Pandey, L.D. Malviya , Vineet Sharma “ Comparative Study of LMS and NLMS Algorithms in Adaptive Equalizer”, International Journal of Engineering Research and Applications (IJERA), Vol. 2, Issue 3, May-Jun 2012, pp.1584-1587. [10] Radhika Chinaboina, 1D.S.Ramkiran, 2Habibulla Khan, 1M.Usha, 1B.T.P.Madhav, 1K.Phani Srinivas & 1G.V.Ganesh,“Adaptive Algorithms For Acoustic Echo Cancellation In Speech Processing”, www.arpapress.com/Volumes/Vol7Issue1/IJRRAS, April 2011 [11] S. Haykin, Adaptive Filter Theory, Fourth Edition, Prentice Hall, Upper Saddle River, N.J., 2002. 1677 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 EFFICIENT DETECTION IN DDOS ATTACK FOR TOPOLOGY GRAPH DEPENDENT PERFORMANCE IN PPM LARGE SCALE IPTRACEBACK S.Abarna1, R.Padmapriya2 1 Mphil Scholar, 2Assistant Professor, Department of Computer Science, NadarSaraswathi College (India) ABSTRACT A variety of schemes based on the technique of Probabilistic Packet Marking (PPM) have been proposed to identify Distributed Denial of Service (DDoS) attack traffic sources by IP trace back. These PPM-based schemes provide a way to reconstruct the attack graph - the network path taken by the attack traffic - hence identifying its sources. Despite the large amount of research in this area, the influence of the underlying topology on the performance of PPM-based schemes remains an open issue. Distributed Denial-of-Service (DDoS) attacks are a critical threat to the Internet. However, the memory less feature of the Internet routing mechanisms makes it extremely hard to trace back to the source of these attacks. In this paper, we propose a novel trace back method for DDoS attacks that is based on entropy variations between normal and DDoS attack traffic, which is fundamentally different from commonly used packet marking techniques.we identify five network-dependent factors that affect different PPM-based schemes uniquely giving rise to a variation in and discrepancy between scheme performances from one network to another. Using simulation, we also show the collective effect of these factors on the performance of selected schemes in an extensive set of 60 Internet-like networks. We find that scheme performance is dependent on the network on which it is implemented. We show how each of these factors contributes to a discrepancy in scheme performance in large scale networks. This discrepancy is exhibited independent of similarities or differences in the underlying models of the networks. I. INTRODUCTION Internet Protocol (IP) trace back is a technique for identi- fying the sources of Distributed Denial of Service (DDoS) attacks from its traffic . One approach to implementing IP trace back ensures that the routers embed their identity in packets randomly selected from all the packets they process . In the event of an attack, the victim uses the packets that contain router identities to construct an attack graph. The attack graph is a representation of the routers and links that the attack packets traversed from the attacker(s) to the victim. This IP trace back type is called probabilistic packet marking (PPM) and is implemented by PPM-based schemes. A lot of intensive research has gone into designing PPM- based schemes that are computationally more efficient and robust than the original PPM. However, little work has gone into identifying network-dependent factors that affect the performance of PPM-based schemes in large-scale networks. In fact, most simulations are carried out on disparate tree- structured topologies and the analytical models derived from these topologies are used to predict the performance of the schemes when deployed in a large-scale network such as the Internet . However, since the schemes are implemented on disparate networks, it is difficult to directly compare the performance of 1678 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 different schemes. Furthermore, because typical underlying topologies are tree-structured, it is difficult to make appropriate projections about scheme performance in a large-scale network without implementing the scheme on that network. In this work, we show the influence of network topology on PPM-based scheme performance. We identify three network- dependent factors that affect scheme performance in large- scale networks. These factors include average shortest path length, overlapping of attack paths, and the occurrence of network motifs in attack graphs. Using specific attack graphs, we show the influence of each factor on selected PPM- based schemes. We then use 60 Internet-like networks to show how all the identified factors collectively contribute to the performance of PPM-based schemes in more realistic scenarios. The networks are selected to encompass the variety of mathematical models used by researchers to create networks that adequately describe the structure of the Internet. Results show that PPM-based scheme performance is de- pendent on the network on which it is implemented. In fact, even the order of performance changes from one network to another, i.e. the best performing scheme in one network is not necessarily the best performing scheme in another network. Our results show how the identified factors contribute, both individually and collectively, to the PPM-based schemes’ performance in large scale networks. II.UNDERLYING TOPOLOGIES A variety of underlying topologies have been used to evaluate the performance of PPM-based schemes. Some schemes use a single path single attacker (SP/SA) topology to simulate a Denial of Service (DoS) attack [2], [3], [5], [6], [8]. In these cases, the length of the attack path is different ranging from 3 hops to 32 hops. Other schemes utilize trees as their underlying topologies for simulation. These include binary trees in [4], [3], [10] ranging from 6 hops to 10 hops. Other tree structures such as random tree networks are used in [3], [8]. Yet other schemes utilize internet topology datasets such as traceroute datasets and skitter maps [9]. While these topologies are significantly larger than other topologies (up to 174,409 nodes), they do not easily lend themselves to simulations and, consequently, comparisons with other schemes. Fig 2. Binary Tree Topology III. SIMULATION STUDY The 60 networks that are considered in this study and some of their properties. These properties include the setup properties such as the underlying model, and appropriate settings required to build each specific network. 1679 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 Additionally, network specific properties, e.g. average shortest path length, and network motif IDs are shown. Each of the networks consists of 1000 nodes representing routers in a network, all of which employ the marking schemes. One of the nodes is selected to be the victim and 50 other nodes are randomly selected to be the attackers. Constant Bit Rate (CBR) sources of traffic are implemented at the attackers, and the convergence time for the entire attack graph is measured in packets. This simulation is executed 200 times for each network and each marking scheme. Fig 3. (a) Number of Network Attack Notes Fig 3. (b) Constant Bit Rate IV. OVERLAPPING OF ATTACK PATHS In this subsection, we show how the level of overlap between two attack paths affects the schemes’ convergence times. We consider a Y-shaped attack graph linking attackers A1 and A2 to victim V (cf. Fig. 2). While keeping each attack path equal and constant, we vary the amount of overlap between the attack paths and observe how the convergence times of PPM, TMS, and PBS are affected. Fig. 6 shows the observed results from this investigation. The results show that there is a general reduction in con- vergence times for all considered schemes as the percentage overlap is increased. Despite the general reduction for all 3 considered schemes, the level of overlap affects each scheme uniquely. For example, the results show that PPM and TMS are relatively unaffected by low amounts of overlap, i.e. 0-20% while PBS exhibits a reduction in convergence times in the same overlap range. However, further increase in percentage overlap causes a drastic decrease in the convergence time of TMS such that by 60%-70% TMS has lower conver- gence times than both PBS and PPM. These results show three things: Firstly, larger amounts of overlapping attack paths translates to reduced convergence times; Secondly, low amounts of overlapping attack paths affects PBS more than PPM and TMS; and thirdly, medium amounts of overlapping cause a drastic reduction in TMS convergence times. In the context of a larger network, these results mean that even for long path lengths, the existence of common and therefore overlapping attack paths translates to reduced convergence times for TMS and PBS more than it does for PPM. 1680 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 Fig 4. (a) TMS convergence times Fig 4.(b) PBS V. MAXIMUM ENTROPY RANDOM WALK Instead of using GRW, Burda et al. [15] introduce the construction of Maximum Entropy Random Walk (MERW), where the transition matrix is defined by the entries pi,j as follows. pi,j =ai,j λ ψjψi Fig 5.Entropy Random Walk The term λ is the largest eigenvalue of A and ψ its corresponding normalized eigenvector withi ψ2 i =1. Thus, the transition probabilities of the random walk process are similar to the eigenvector centrality of the nodes, which is regarded as one way of describing the influence of a node within the topology. The same way of defining entropy within a network and setting the transition probabilities accordingly to maximize entropy were also discussed by Demetrius and Manke in [17]. They further establish a relationship between entropy and the robustness of the average shortest path length since networks that have a higher entropy are also more robust toward removal of nodes. The stationary distribution of finding a query packet at node i with MERW and its entropy rate are then according to [15] as shown in Eqns. (6) and (7).π∗ i = ψ2 i (6) SMERW = logλ (7)Unfortunately, the definition of the transition probabilities in Eqn. (5) requires knowledge of the largest eigenvalue of the adjacency matrix and its corresponding eigenvector. This can only be determined if the topology is fully known and is usually not very practical, especially in large networks. However, it was shown by Sinatra et al. [16] that the maximum entropy random walk can be constructed only with limited and local information based on the degrees of the first and second hop neighbors of each node. 1681 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 5.1 Average Number of Hops The average number of hops for a query packet indicates the speed of search within the network. Figure 3 shows the average hop count until finding one of the destinations for the three different methods with R = 50destination nodes and the network size N varying from 200 to 1000. The results for the scale-free topology are shown in Fig. 3(a) and for the small-world topology in Fig. 3(b). Fig 5.(a) Scale-Free Topoloy Fig5(b) Smal –world topology VI. CONCLUSION AND FUTURE WORK In this paper we studied the feasibility of applying a random walk query search in a data-centric network under random and complex topologies. For suitable topologies, random walk does not necessarily perform much worse than the commonly used flooding mechanism. On the contrary, since only a single path is followed during the query search, fewer nodes are involved in the dissemination process, which leads to a lower consumption of energy. We have seen that a maximum entropy random walk improves the general random walk in performance by counteracting the irregularities in topology to balance the reachability probability of the destination nodes. Furthermore, only a small number of destination nodes in the network is sufficient to provide replicas of the desired content for achieving performance compared to flooding for both small world and scale-free topologies. REFERENCES [1] C. Intanagonwiwat, R. Govindan, and D. Estrin, “Directed diffusion: a scalable and robust communication paradigm for sensor networks”, in Proc. 6th Annu. Int. Conf. on Mobile Computing and Networking (MobiCom ’00), pp. 56–56, Boston, MA, 2000. [2] I. Stojmenovic´ and S. Olariu, “Data-centric protocols for wireless sensor networks”, in Handbook of sensor networks: algorithms and architectures, I. Stojmenovic´ (Ed.), John Wiley & Sons, 2005. [3] V. Jacobson, D. K. Smetters, J. D. Thornton, M. Plass, N. Briggs, and R. Braynard. 2012. “Networking named content”. Commun. ACM, vol. 55, no. 1, pp. 117–124, Jan. 2012. [4] J. Choi, J. Han, E. Cho, T. Kwon, and Y. Choi, “A survey on content-oriented networking for efficient content delivery”, IEEE Com- mun. Mag., vol. 49, no. 3, pp. 121–127, March 2011. [5] A. Mainwaring, D. Culler, J. Polastre, R. Szewczyk, and J. Anderson, “Wireless sensor networks for habitat monitoring”, in Proc. 1st ACM Int. Workshop on Wireless Sensor Networks and Applications (WSNA ’02), pp. 88–97, Atlanta, GA, September 2002. 1682 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 DISTRIBUTED RESTRAINING MALWARE PROPAGATION IN MOBILE SINK ROUTING FOR WIRELESS SENSOR NETWORKS M.Sindhuja1, A.Komathi2 1 Mphil Scholar, Department of CS & IT, Nadar Saraswathi College of Arts and Science, (India) 2 Department of CS & IT, Nadar Saraswathi College of Arts and Science, (India) ABSTRACT Advances in wireless sensor network (WSN) technology has provided the availability of small and low-cost sensor nodes with capability of sensing various types of physical and environmental conditions, data processing, and wireless communication. Variety of sensing capabilities results in profusion of application areas. However, the characteristics of wireless sensor networks require more effective methods for data forwarding and processing. In WSN, the sensor nodes have a limited transmission range, and their processing and storage capabilities as well as their energy resources are also limited. Routing protocols for wireless sensor networks are responsible for maintaining the routes in the network and have to ensure reliable multi-hop communication under these conditions. In this paper, we give a survey of routing protocols for Wireless Sensor Network and compare their strengths and limitations. Keywords: Wireless Sensor Networks, Routing Protocols, Cluster Head I. INTRODUCTION Wireless sensor network (WSN) is widely considered as one of the most important technologies for the twentyfirst century [1]. In the past decades, it has received tremendous attention from both academia and industry all over the world. A WSN typically consists of a large number of low-cost, low-power, and multifunctional wireless sensor nodes, with sensing, wireless communications and computation capabilities [2,3]. These sensor nodes communicate over short distance via a wireless medium and collaborate to accomplish a common task, for example, environment monitoring, military surveillance, and industrial process control [4]. The basic philosophy behind WSNs is that, while the capability of each individual sensor node is limited, the aggregate power of the entire network is sufficient for the required mission. Although several defense mechanisms [6, 7] have been pro- posed in the literature over the last few years, little work has been done to demonstrate howvulnerable, in terms of data confidentiality and network availability, these networks are. Motivated by this unexplored security aspect, wedemon- strate an attack tool that can be useful not only in highlight- ing the importance of defending sensor networkapplications against attacks but also in studying the effects of these attacks on the sensor network itself. This in turn can lead to the development of more secure applications and better detection/prevention mechanisms. 1683 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 II. ATTACK TOOL ARCHITECTURE OVERVIEW The attack tool is based on an intelligent component-based system. The hosted components are capable of monitoring any neighborhood traffic, decoding and logging overheard packets, constructing specially crafted messages and launch- ing a number of attacks. Its core functionality is based on three main conceptual modules, as depicted in Figure 1: • A network sniffer for passive monitoring and logging of radio packets. Any network traffic analysis or packet decoding can be done either in real time or offline through the implemented packet description database. • A network attack tool that provides a number of actions for compromising a sensor network’s security profile. It contains a data stream framework for constructing specially crafted packets that are transmitted by the attack launcher throughout the duration of an attack. • A network visualization component that visualizes and displays the neighborhood topology, network traffic, node states and status of any performed attack. The key design goal of this tool is its wide applicability; it should support passive inspection and compromise of a wide variety of sensor network protocols and applications. Figure 1: Attack Tool Architecture Layout. III. NETWORK CONFIDENTIALITY THREATS AND WIRELESS ATTACKS In wireless networking the overall security objectives remain the same as with wired networks: preserving confidentiality, ensuring integrity, and maintaining availability of information. Thus, identifying risks to sensor networks confiden- tiality posed by theavailability of transactional data is ex- tremelyvital.In an attempt to identify network confidentiality threats, we enhanced our attack tool with a network sniffer for overhear- ing network traffic (Section 3.1). In that way an adversary can process transmitted packets in order to extract vital in- formation such as node IDs or traffic data. Our assertion is that traffic analysis can provide more information about a network’snodes and usage than simply decoding any data packet contents. The presented tool can use carrier frequency to launch a side-channel attack [18] in an attempt to identify the net- work’s sensor hardware platform. An adversary could use either a spectrum analyzer or different sensor hardware in combination with our tool in order to detect the current communication frequency. Once the adversary discovers it, she can determine the hardware used and, thus, exploit all the protocol vulnerabilities arising from this specific platform. This tool can also compromise a network’s confidentiality by monitoring the rate and size of any transmitted/received messages. Specifically, the message rate can reveal infor- mation about the network application and the frequency of monitored events. This constitutes a severe threat since for some sensor applications, like health monitoring, it can lead to a violation of user’s privacy. Furthermore, an adversary can examine the rate at which she overhears messages com- ing from a neighborhood and estimate the distance to 1684 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 the sensed event. Research has shown that the message recep- tion rate increases when the distance to the event reporting node decreases. Fig 2. Transactions Vector 3.1 Network Sniffer Component The network sniffer relies on packets that are overheard in a sensor’s node neighborhood. It captures them and logs them for later analysis. Conceptually the sniffer consists of a Local Packet Monitoring module for gathering audit data to be forwarded, over its serial port, to the Packet Storage module for logging at the attached host. This allows offline analysis, through the Packet Description Database, in order to extract vital network information such as node IDs, traffic data or used protocol versions. Essentially, the sniffer enables the construction of a directed graph of all neighboring nodes. Overheard packets flow along the edges of the graph, as Fig 3. Wireless Network Sniffer Audit data consist of the communication activities within the sniffer’s radio range. Such data can be collected by lis- tening promiscuously to neighboring nodes’ transmissions. By promiscuously we mean that when a node is within ra- dio range, the local packet monitoring module can overhear communications originating from that node. Once captured by the radio, all packets are timestamped in order to facili- tate subsequent time-based analysis. Timestamping is per- formed the moment the packet is received by the network sniffer. 1685 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 3.2 Network Attack Tool Component This component core functionality is to provide a number of actions for compromising the sensor network’s security pro- file. After gathering audit data that are used by the network sniffer to extract vital information and identify the used sen- sor hardware platform and underlying protocols, a user can start launching a number of attacks Fig 3.2 (a) Hardware Platform Protocoal Fig 3.2 (b) List of Supported at- Tacks Can be Found The resulting network information stream from the packet decoder is fed to the Data Stream Framework of the attack tool component. This data stream processor uses the identi- fied carrier frequency, message size and routing information as its configuration record. All these network characteristics are essential since they are used as the basis for any specially crafted message required by the Attack Launcher. IV. IMPLEMENTED ATTACKS & ACTIONS Many sensor network deployments are quite simple, and for this reason they can be even more susceptible to attacks. What makes it particularly easy for attackers is the fact that most protocols are not designed having security threats in mind. As a consequence, they rarely include securityprotection and little or no effort is usually required from the side of an adversary to perform an attack. So, it is very important to study realistic attacker models and evaluate their practicality and effectiveness through a tool as the one presented in this work. 1686 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 The nature of wireless network communications opens the way to four basic attacks: Interception, Alteration, Disrup- tion and Code or Packet Injection [5]. Most network layer attacks against such networks fall into one of these cate- gories. Our attack tool (in its current version) gives the user the opportunity to perform, in addition to eavesdrop- ping and sniffing, the following actions: V. CONCLUSIONS In this paper, we have identified some of the sensor net- works vulnerabilities that can be exploited by an attacker for launching various kinds of attacks. We have demonstrated the practicality of these attacks by building an attack tool for compromising the network’s confidentiality and function- ality. The results of this work serve a three-fold purpose: to reveal the vulnerabilities of such networks, to study the ef- fects of severe attacks on the network itself and to motivate a better design of security protocols that can make them more resilient to adversaries. Wireless sensor network security is an important research direction and tools like the current one may be used in coming up with even more attractive solutions for defending these types of networks. REFERENCES [1] A. Mainwaring, D. Culler, J. Polastre, R. Szewczyk, and J. Anderson. Wireless sensor networks for habitat monitoring. In WSNA ’02: Proceedings of the 1st ACM international workshop on Wireless sensor networks and applications, pages 88–97, New York, NY, USA, 2002. ACM. [2] J. Tateson, C. Roadknight, A. Gonzalez, T. Khan, S. Fitz, I. Henning, N. Boyd, C. Vincent, and I. Marshall. Real World Issues in Deploying a Wireless Sensor Network for oceanography. In Workshop on Real-World Wireless Sensor Networks REALWSN’05, Stockholm, Sweden, June 2011. [3] D. Trossen, D. Pavel, G. Platt, J. Wall, P. Valencia, C. A. Graves, M. S. Zamarripa, V. M. Gonzalez, J. Favela, E. Livquist, and Z. Kulcs Sensor networks, wearable computing, and healthcare applications. IEEE Pervasive Computing, 6:58–61, 2007. [4] A. Becher, Z. Benenson, and M. Dornseif. Tampering with motes: Real-world physical attacks on wireless sensor networks. In J. A. Clark, R. F. Paige, F. Polack, and P. J. Brooke, editors, SPC, volume 3934 of Lecture Notes in Computer Science, pages 104–118. Springer, 2006. [5] C. Karlof and D. Wagner. Secure routing in wireless sensor networks: Attacks and countermeasures. AdHoc Networks Journal, 1(2–3):293–315, September 2003. 1687 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 PRESERVING CLOUD CONSISTENCY USING CAAS MODEL P. Malaimari Suganya1 , R.Padmapriya2 1 Mphil Scholar, 2Assistant Professor, Department Of Computer Science , Nadar Saraswathi College (India) ABSTRACT Cloud storage services are commercially more popular due to their amount of advantages. Most of the cloud service provider provides services like infrastructure management, data storage services on 24/7 through any devices at anywhere. To provide this ubiquitous always on service most of the cloud service provider (CSP) maintains each piece of data on geographically distributed servers. The main key problem with this technique is that, it is very expensive and some to fail to provide required consistency of service. To overcome this problem, we propose to use a new approach of service (i.e. Consistency as a Service(CaaS)) this paper, firstly concentrate on a consistency as a service (CaaS) model, which has a large data cloud and multiple small audit clouds. In the CaaS model, a data cloud is formed by a CSP, and a group of users form an audit cloud that can verify whether the data cloud provides the promised level of consistency i.e. quality of service or not, for that make use of two-level auditing strategy which require loosely synchronized clock for ordering operations in an audit cloud. Then perform global auditing by global trace of operations through randomly electing an auditor from an audit cloud. Finally, use a heuristic auditing strategy(HAS) to display as many violations as possible. Keywords: Cloud Storage Systems, Consistency As A Service (Caas), Two-Level Auditing and Heuristicauditing Strategy I. INTRODUCTION Clouds computing is become more popular as it provides guaranteed services like data storage, virtualized infrastructure etc. e.g. Amazon,SimpleDB etc. By using the cloud services, the customers or user can access data stored in a cloud anytime and at anywhere using any device, and customer ensure about less capital investment. To provide promised always on 24/7 access, the cloud service provider (CSP) stores data replicason multiple geographically distributed servers. The main drawback of using the replication technique is it is very expensive to achieve strong consistency, and user is ensured to see the latest updates. Many CSPs (e.g., Amazon S3) provide only eventual i.e. updates are visible definitely but not immediately. E.g. Domain name system (DNS), but the eventual consistency is not interesting for all applications and which require strong consistency. Some applications like social networking sites require causal i.e. strong consistency. Thus the different applications require different level of consistency. We propose novel consistency as a service (CAAS) model. The CaaS model consists of, A large data cloud formed by CSP and multiple audit clouds formed by group of users worked on project or document that can check whether the data cloud provide a promised level of consistency or not. Two-level auditing structure which require only a loosely synchronized clock for ordering operation in an audit cloud then perform global auditing with a global trace of operations periodically an auditor 1688 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 is elected from an audit cloud. Local auditing is concentrate on monotonic-read and read-your-write consistencies, which can be performed by an online light-weight algorithm while Global auditing focuses on causal consistency, in which construct a directed graph. If the constructed graph is a directed acyclic graph also called as precedence graph, we claim that causal consistency is preserved. We determine the severity of violations by two metrics for the CaaS model: commonality of violations and staleness of the value of a read, as in. Finally, we propose a heuristic auditing strategy (HAS) which adds appropriate reads to display as many violations as possible to determine cloud consistency and also actual cost per transaction. II. DESCRIPTION This section consist of three models i. e. consistency as a service (CaaS) model, user operation table (UOT) with which each user records his operations and two-level auditing structure. 2.1 Consistency as a Service (CAAS) Model An audit cloud consists of a group of users that work together on a job, e.g., a document or a program. We consider that each user in the audit cloud is identified by a unique ID. Before assigning job to the data cloud, an audit cloud and the data cloud will engage with a service level agreement (SLA), which demands the promised level of consistency should be provided by the data cloud. The audit cloud exists to verify whether the data cloud violates the SLA or not, and to analyze the severity of violations. 2.2 User Operation Table (UOT) Each user maintains his own User Operation Table (UOT) for recording his trace of operations. Each record in the UOT is described by elements like Operation, logical vector, and physical vector. While issuing an operation, a user from an audit cloud will record his operation in UOT, as well as his current logical vector and physical vector. Each user will maintain a logical vector and a physical vector to track the logical and physical time when an operation happens, respectively. 2.3 Two-Level Auditing Structure 2.3.1 Local Auditing Each user independently performs local auditing with his UOT with two consistencies; Monotonic-read consistency, which requires that a user must read either a new value or same value Read-your’s-write consistency, which require a user, always read his latest update. 2.3.2 Global Auditing Global auditing is performed by global trace of operations of all users operations with following consistency Causal Consistency Causal consistency writes that are causally related must be seen by all process in the same order and concurrent writes may be seen in a different order on different machine. 1689 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 ISBN: 978-81-931039-3-7 Fig:1 Consistency as a Service Model 2.4 Heuristic Auditing Strategy From the auditing process it is clear that only reads can display violations by their values. Therefore, the basic idea behind the heuristic auditing strategy (HAS) is to add exact reads for displaying as many violations as possible and call these additional reads as auditing reads. Under the CaaS model, consistency becomes a part of the Service Level Agreement and the users can get something from the CSP, by displaying consistency violations and determine the severity of the violations. The CaaS model will help both the CSP and the users adopt consistency as an important aspect of cloud services. III. CONCLUSION In this paper, we argued that strong consistency requirements should be adopted only for data objects crucial for application correctness, otherwise weaker forms of data consistency should be adopted. We presented aconsistency as a service (CaaS) model and a two-level auditing structure that helps users to verify whether the cloud service provider (CSP) is providing the promised consistency, and to quantify the severity of the violations, if any. With the CaaS model, the users can assess the quality of cloud services and choose a right CSP among various candidates, e.g., the least expensive one that still provides adequate consistency for the users’ applications. REFERENCE [1] in Liu, Guojun Wang, IEEE, Member, and Jie Wu, IEEE Fellow “Consistency as a service: Auditing cloud consistency ", IEEE Vol 11 No.1, July 2014. [2] A. TANENBAUM AND M. VAN STEEN, Distributed Systems: Principles andParadigms. Prentice Hall PTR, 2002 1690 | P a g e International Conference on Emerging Trends in Technology, Science and Upcoming Research in Computer Science DAVIM, Faridabad, 25th April, 2015 [3] ISBN: 978-81-931039-3-7 W. VOGELS, “Data access patterns in the Amazon.com technology platform,” in Proc. 2007 VLDB. [4] E. BREWER, “Towards robust distributed systems,” in Proc. 2000 ACM PODC. [5] Pushing the CAP: strategies for consistency and availability,” Computer, vol. 45, no.2, 2012 [6] E. ANDERSON, X. LI, M. SHAH, J. TUCEK, AND J. WYLIE, What consistency does your key-value store actually provide,” in Proc. 2010 USENIX HotDep. [7] W. GOLAB, X. LI, AND M. SHAH, “Analyzing consistency properties for fun and profit,” in Proc. 2011 ACM PODC 1691 | P a g e