Edge Detection Based on Improved Sobel Operator

Total Page:16

File Type:pdf, Size:1020Kb

Edge Detection Based on Improved Sobel Operator Advances in Computer Science Research, (ACSR), volume 52 2016 International Conference on Computer Engineering and Information Systems (CEIS-16) Edge Detection Based on Improved Sobel Operator Chao-Chao Zhang, Jian-Dong Fang Inner Mongolia University of Technology Electric Engineering College, Hohhot 010080, China E-mail: [email protected], [email protected] Abstract-Sobel algorithm is an important method of image edge many respects, Sobel operators are superior to others. detection. Comparing the Sobel operator with several other However, the classical Sobel operator also has some edge detection operators used frequently and making a further problems. The Sobel operator is sensitive only to the study on the classical Sobel operator, the advantages of Sobel vertical direction and the horizontal direction. However, the operator are its fast detection speed, meanwhile , it has an image information is not limited to the horizontal and effect on smoothing and suppressing noise. Also, Sobel vertical direction; it can make part of the image information operator has a good effect on edge detection. Although Sobel lose. In this paper, a new improved operator is proposed operator has advantages in many aspects, it exists some based on the sobel operator. On the traditional Sobel problems: the Sobel operator is a kind of edge detection in operator, the 45 degrees and 135 degrees 2 direction horizontal and vertical direction, so it neglects edge points in other directions. It can not achieve a true detection for the templates are added to realize multi- directional image points on image edge. In this paper, the algorithm is based on acquisition. Then calculate the threshold by using the Otsu the Sobel operator, an increase of 45 degrees and 135 degrees 2 method and refine the detected rough edges by using direction template, while the main edge of the oblique, binarization method to achieve the results of image edge re-assigned the weight of the operator template. At the same detection. Edge detection effect can be achieved better by time, in order to achieve an effect of detection, binarization using Matlab simulation method. method is used to make an edge thinning for detected image. A comparison table of various operators is shown in According to simulation experiments, they show this method is Table 1: simple and feasible, and the detective result is more concrete and abundant than traditional Sobel edge detection. Some II. THE COMPARISON OF SEVERAL problems are improved, such as traditional Sobel edge is rough TRADITIONAL OPERATORS and detection is incomplete. • Roberts operator: did not pass smooth calculation, so it is more sensitive to the noise. Keywords-sobel operator; binarization; edge detection; edge thinning • Prewitt operator and Sobel operator: edge extraction effect is almost the same. Prewitt operator is an I. INTRODUCTION average filter, Sobel operator is a weighted average filter; they have better detection effect on images In digital image processing, edge feature is one of the which have low gray level noise, but the detection important features of the image, and it is an important part effect is not obvious. of image processing, pattern recognition and computer • vision. The results of image edge detection affect the further LOG operator: detecting edge by using two order image processing and pattern recognition directly. In recent derivatives zero crossing edge method. Smoothing decades, the image edge detection technology has become effect is more significant, removing effect of noise an important research topic in digital image processing is better, but the loss of image details is larger, the technology. With the development of science and edge precision is lower. So there is a contradiction technology, researchers have proposed some methods for between positioning edge accuracy and eliminating image edge detection and evaluation of edge detection. At noise level. the same time, these edge detection techniques are applied III. THE PRINCIPLES OF THE TRADITIONAL to the field of computer vision and pattern recognition, SOBEL OPERATOR which makes the application of edge detection technology more and more widely [1]. Over the years, image Sobel cross-differential operator, because of its role in segmentation has been attracting more and more attention. the template is small, the relative calculation is also small, Thousands of segmentation algorithms have been put thus the image details of the contour can be obtained. But forward. They can be divided into bit threshold method, because the template size is even, the pending pixel cannot edge detection method and regional growth method [2]. be placed in the center position of the template, or the Edge detection method includes: first order differential edge results will have half pixel fault. Sobel differential operator detection operator includes Roberts operator, Prewitt is a directional differential operator in a basis of odd size operator, LOG operator and Sobel operator.In contrast, in template. The expressions of formula as follow (3.1), (3.2): Copyright © 2016, the Authors. Published by Atlantis Press. 129 This is an open access article under the CC BY-NC license (http://creativecommons.org/licenses/by-nc/4.0/). Advances in Computer Science Research, (ACSR), volume 52 G( i , j )= f[][] i − 1, j + 1 + 2 × f i , j + 1 idea of Sobel operator algorithm: because the edge of the x image is located at the place in which the brightness +f[][] i +1, j + 1 − f i − 1, j − 1 (3.1) changes significantly, therefore in the neighborhood of the pixel gray value of pixels exceeds a set threshold depending −2 ×f[][] i , j − 1 − f i + 1, j − 1 on the specific steps for the edge [3]. The specific steps of Sobel operator algorithm are as follows [4]: = + − + × + • Moving the horizontal and vertical direction Gy () i, j f[][] i 1, j 1 2 f i 1, j templates from left to right, from top to bottom, and moving from one pixel to another, and the center of (3.2) +f[][] i +1, j + 1 − f i − 1, j − 1 the template corresponds to the corresponding −2 ×f[][] i − 1, j − f i − 1, j + 1 pixels in the image. • Multiplying the coefficients in the template with the corresponding pixel values in the image. The convolution template of the Sobel operator is • The gradient value is calculated as a new gray value expressed as the formula (3.3) and (3.4): by using 2 convolution values. • Select the appropriate threshold TH, if the new pixel −1 − 2 − 1 value ≥ TH, the pixel points can be regarded as = image edge points. Gx 0 0 0 (3.3) 1 2 1 IV. THE IMPROVEMENT OF THE TRADITIONAL SOBEL OPERATOR −1 0 1 Adding 45 degrees and 135 degrees direction template in a basis of the traditional Sobel operator, the direction G = − 2 0 2 (3.4) template are changed into four directions: the horizontal and y vertical direction and 45 degrees and 135 degrees direction. −1 0 1 It improved the weights of new template in the direction of oblique edge. Specific details are as follows: The calculating steps of Sobel operator: first, the edge According to the calculation of the four template detection image is divided into matrix form (3.5): −1 − 2 − 1 - 2 -1 0 0 0 0 -1 0 1 1 2 1 0 1 2 f1()()() x−1, y − 1 f1 x − ,1 y f1 x−1, y + 1 horizontal direction 45 degree direction A = f1()()() x, y− 1 f1 x, y f1 x, y + 1 (3.5) −1 0 1 0 1 2 f1()()() x+1, y − 1 f1 x + ,1 y f1 x+1, y + 1 − 2 0 2 -1 0 1 Multiply horizontal direction by vertical direction of the −1 0 1 - 2 -1 0 template and then multiply the vertical direction by vertical direction 135 degree direction horizontal direction of the template, Fx=Gx.*A, Fx=Gy.*A. directions and the calculating an image point by point, the gradient size calculation, as shown in the formula (3.6): maximum value is regarded as the pixel gray values. According to the threshold setting, edge point is determined. 2 2 The template maximum value corresponding to represents GG= +G (3.6) the direction of edge direction of the pixel. x y A. Edge Thinning Processing The formula for calculating the gradient direction is shown The edge of the image contains a lot of image in the formula (3.7): information, by the direction of the traditional Sobel operator improvement, only the gradient information. The G edge appears where the gradient value is large. To extract y θ = arctan (3.7) the edge of the image, a threshold must be set and then Gx binarized. In general, the gradient image obtained by the gradient operator has a thicker edge. If the gradient image When θ is equal to zero, there is a vertical edge on threshold is set directly to binarization, it is difficult to find behalf of the image, the left side is dark er than right side. a suitable threshold, so that the detected edge can meet the The Sobel operator introduced the weighted local requirements. This is not easy to edge processing and edge average, it can not only affect the image edge detection but feature extraction and other post-processing.Therefore, it is also suppress noise further, but the edge is wider. The basis necessary to refine the edge of the gradient before 130 Advances in Computer Science Research, (ACSR), volume 52 binarizing the image gradient [5]. separate Sobel operator for image edge detection, and then The refinement process is to find the pixel (a, b) Max a use the algorithm for image edge detection; make a maximum value in the neighborhood of (a, b), and set the comparison between them, Figure 1and4 shows the original, threshold locally according to Max (a, b).
Recommended publications
  • Edge Detection of Noisy Images Using 2-D Discrete Wavelet Transform Venkata Ravikiran Chaganti
    Florida State University Libraries Electronic Theses, Treatises and Dissertations The Graduate School 2005 Edge Detection of Noisy Images Using 2-D Discrete Wavelet Transform Venkata Ravikiran Chaganti Follow this and additional works at the FSU Digital Library. For more information, please contact [email protected] THE FLORIDA STATE UNIVERSITY FAMU-FSU COLLEGE OF ENGINEERING EDGE DETECTION OF NOISY IMAGES USING 2-D DISCRETE WAVELET TRANSFORM BY VENKATA RAVIKIRAN CHAGANTI A thesis submitted to the Department of Electrical Engineering in partial fulfillment of the requirements for the degree of Master of Science Degree Awarded: Spring Semester, 2005 The members of the committee approve the thesis of Venkata R. Chaganti th defended on April 11 , 2005. __________________________________________ Simon Y. Foo Professor Directing Thesis __________________________________________ Anke Meyer-Baese Committee Member __________________________________________ Rodney Roberts Committee Member Approved: ________________________________________________________________________ Leonard J. Tung, Chair, Department of Electrical and Computer Engineering Ching-Jen Chen, Dean, FAMU-FSU College of Engineering The office of Graduate Studies has verified and approved the above named committee members. ii Dedicate to My Father late Dr.Rama Rao, Mother, Brother and Sister-in-law without whom this would never have been possible iii ACKNOWLEDGEMENTS I thank my thesis advisor, Dr.Simon Foo, for his help, advice and guidance during my M.S and my thesis. I also thank Dr.Anke Meyer-Baese and Dr. Rodney Roberts for serving on my thesis committee. I would like to thank my family for their constant support and encouragement during the course of my studies. I would like to acknowledge support from the Department of Electrical Engineering, FAMU-FSU College of Engineering.
    [Show full text]
  • EECS 442 Computer Vision: Homework 2
    EECS 442 Computer Vision: Homework 2 Instructions • This homework is due at 11:59:59 p.m. on Friday February 26th, 2021. • The submission includes two parts: 1. To Canvas: submit a zip file of all of your code. We have indicated questions where you have to do something in code in red. Your zip file should contain a single directory which has the same name as your uniqname. If I (David, uniqname fouhey) were submitting my code, the zip file should contain a single folder fouhey/ containing all required files. What should I submit? At the end of the homework, there is a canvas submission checklist provided. We provide a script that validates the submission format here. If we don’t ask you for it, you don’t need to submit it; while you should clean up the directory, don’t panic about having an extra file or two. 2. To Gradescope: submit a pdf file as your write-up, including your answers to all the questions and key choices you made. We have indicated questions where you have to do something in the report in blue. You might like to combine several files to make a submission. Here is an example online link for combining multiple PDF files: https://combinepdf.com/. The write-up must be an electronic version. No handwriting, including plotting questions. LATEX is recommended but not mandatory. Python Environment We are using Python 3.7 for this course. You can find references for the Python standard library here: https://docs.python.org/3.7/library/index.html.
    [Show full text]
  • Exploiting Information Theory for Filtering the Kadir Scale-Saliency Detector
    Introduction Method Experiments Conclusions Exploiting Information Theory for Filtering the Kadir Scale-Saliency Detector P. Suau and F. Escolano {pablo,sco}@dccia.ua.es Robot Vision Group University of Alicante, Spain June 7th, 2007 P. Suau and F. Escolano Bayesian filter for the Kadir scale-saliency detector 1 / 21 IBPRIA 2007 Introduction Method Experiments Conclusions Outline 1 Introduction 2 Method Entropy analysis through scale space Bayesian filtering Chernoff Information and threshold estimation Bayesian scale-saliency filtering algorithm Bayesian scale-saliency filtering algorithm 3 Experiments Visual Geometry Group database 4 Conclusions P. Suau and F. Escolano Bayesian filter for the Kadir scale-saliency detector 2 / 21 IBPRIA 2007 Introduction Method Experiments Conclusions Outline 1 Introduction 2 Method Entropy analysis through scale space Bayesian filtering Chernoff Information and threshold estimation Bayesian scale-saliency filtering algorithm Bayesian scale-saliency filtering algorithm 3 Experiments Visual Geometry Group database 4 Conclusions P. Suau and F. Escolano Bayesian filter for the Kadir scale-saliency detector 3 / 21 IBPRIA 2007 Introduction Method Experiments Conclusions Local feature detectors Feature extraction is a basic step in many computer vision tasks Kadir and Brady scale-saliency Salient features over a narrow range of scales Computational bottleneck (all pixels, all scales) Applied to robot global localization → we need real time feature extraction P. Suau and F. Escolano Bayesian filter for the Kadir scale-saliency detector 4 / 21 IBPRIA 2007 Introduction Method Experiments Conclusions Salient features X HD(s, x) = − Pd,s,x log2Pd,s,x d∈D Kadir and Brady algorithm (2001): most salient features between scales smin and smax P.
    [Show full text]
  • Image Segmentation Based on Sobel Edge Detection Yuqin Yao1,A
    5th International Conference on Advanced Materials and Computer Science (ICAMCS 2016) Image Segmentation Based on Sobel Edge Detection Yuqin Yao 1,a 1 Chengdu University of Information Technology, Chengdu, 610225, China a email: [email protected] Keywords: MM-sobel, edge detection, mathematical morphology, image segmentation Abstract. This paper aiming at the digital image processing, the system research to add salt and pepper noise, digital morphological preprocessing, image filtering noise reduction based on the MM-sobel edge detection and region growing for edge detection. System in this paper, the related knowledge, and application in various fields and studied and fully unifies in together, the four finished a pair of gray image edge detection is relatively complete algorithm, through the simulation experiment shows that the algorithm for edge detection effect is remarkable, in the case of almost can keep more edge details. Research overview The edge of the image is the most important visual information in an image. Image edge detection is the base of image analysis, image processing, computer vision, pattern recognition and human visual [1]. The ultimate goal is image segmentation; the largest premise is image edge detection. Image edge extraction plays an important role in image processing and machine vision. Proper image detection method is always the research hotspots in digital image processing, there are many methods to achieve edge detection, we expect to find an accurate positioning, strong anti-noise, not false, not missing detection algorithm [2]. The edge is an important feature of an image. Typically, when we take the digital image as input, the image edge is that the gray value of the image is changing radically and discontinuous, in mathematics the point is known as the break point of signal or singular point.
    [Show full text]
  • Topic: 9 Edge and Line Detection
    V N I E R U S E I T H Y DIA/TOIP T O H F G E R D I N B U Topic: 9 Edge and Line Detection Contents: First Order Differentials • Post Processing of Edge Images • Second Order Differentials. • LoG and DoG filters • Models in Images • Least Square Line Fitting • Cartesian and Polar Hough Transform • Mathemactics of Hough Transform • Implementation and Use of Hough Transform • PTIC D O S G IE R L O P U P P A D E S C P I Edge and Line Detection -1- Semester 1 A S R Y TM H ENT of P V N I E R U S E I T H Y DIA/TOIP T O H F G E R D I N B U Edge Detection The aim of all edge detection techniques is to enhance or mark edges and then detect them. All need some type of High-pass filter, which can be viewed as either First or Second order differ- entials. First Order Differentials: In One-Dimension we have f(x) d f(x) dx d f(x) dx We can then detect the edge by a simple threshold of d f (x) > T Edge dx ⇒ PTIC D O S G IE R L O P U P P A D E S C P I Edge and Line Detection -2- Semester 1 A S R Y TM H ENT of P V N I E R U S E I T H Y DIA/TOIP T O H F G E R D I N B U Edge Detection I but in two-dimensions things are more difficult: ∂ f (x,y) ∂ f (x,y) Vertical Edges Horizontal Edges ∂x → ∂y → But we really want to detect edges in all directions.
    [Show full text]
  • Study and Comparison of Different Edge Detectors for Image
    Global Journal of Computer Science and Technology Graphics & Vision Volume 12 Issue 13 Version 1.0 Year 2012 Type: Double Blind Peer Reviewed International Research Journal Publisher: Global Journals Inc. (USA) Online ISSN: 0975-4172 & Print ISSN: 0975-4350 Study and Comparison of Different Edge Detectors for Image Segmentation By Pinaki Pratim Acharjya, Ritaban Das & Dibyendu Ghoshal Bengal Institute of Technology and Management Santiniketan, West Bengal, India Abstract - Edge detection is very important terminology in image processing and for computer vision. Edge detection is in the forefront of image processing for object detection, so it is crucial to have a good understanding of edge detection operators. In the present study, comparative analyses of different edge detection operators in image processing are presented. It has been observed from the present study that the performance of canny edge detection operator is much better then Sobel, Roberts, Prewitt, Zero crossing and LoG (Laplacian of Gaussian) in respect to the image appearance and object boundary localization. The software tool that has been used is MATLAB. Keywords : Edge Detection, Digital Image Processing, Image segmentation. GJCST-F Classification : I.4.6 Study and Comparison of Different Edge Detectors for Image Segmentation Strictly as per the compliance and regulations of: © 2012. Pinaki Pratim Acharjya, Ritaban Das & Dibyendu Ghoshal. This is a research/review paper, distributed under the terms of the Creative Commons Attribution-Noncommercial 3.0 Unported License http://creativecommons.org/licenses/by-nc/3.0/), permitting all non-commercial use, distribution, and reproduction inany medium, provided the original work is properly cited. Study and Comparison of Different Edge Detectors for Image Segmentation Pinaki Pratim Acharjya α, Ritaban Das σ & Dibyendu Ghoshal ρ Abstract - Edge detection is very important terminology in noise the Canny edge detection [12-14] operator has image processing and for computer vision.
    [Show full text]
  • A Comprehensive Analysis of Edge Detectors in Fundus Images for Diabetic Retinopathy Diagnosis
    SSRG International Journal of Computer Science and Engineering (SSRG-IJCSE) – Special Issue ICRTETM March 2019 A Comprehensive Analysis of Edge Detectors in Fundus Images for Diabetic Retinopathy Diagnosis J.Aashikathul Zuberiya#1, M.Nagoor Meeral#2, Dr.S.Shajun Nisha #3 #1M.Phil Scholar, PG & Research Department of Computer Science, Sadakathullah Appa College, Tirunelveli, Tamilnadu, India #2M.Phil Scholar, PG & Research Department of Computer Science, Sadakathullah Appa College, Tirunelveli, Tamilnadu, India #3Assistant Professor & Head, PG & Research Department of Computer Science, Sadakathullah Appa College, Tirunelveli, Tamilnadu, India Abstract severe stage. PDR is an advanced stage in which the Diabetic Retinopathy is an eye disorder that fluids sent by the retina for nourishment trigger the affects the people with diabetics. Diabetes for a growth of new blood vessel that are abnormal and prolonged time damages the blood vessels of retina fragile.Initial stage has no vision problem, but with and affects the vision of a person and leads to time and severity of diabetes it may lead to vision Diabetic retinopathy. The retinal fundus images of the loss.DR can be treated in case of early detection. patients are collected are by capturing the eye with [1][2] digital fundus camera. This captures a photograph of the back of the eye. The raw retinal fundus images are hard to process. To enhance some features and to remove unwanted features Preprocessing is used and followed by feature extraction . This article analyses the extraction of retinal boundaries using various edge detection operators such as Canny, Prewitt, Roberts, Sobel, Laplacian of Gaussian, Kirsch compass mask and Robinson compass mask in fundus images.
    [Show full text]
  • Feature-Based Image Comparison and Its Application in Wireless Visual Sensor Networks
    University of Tennessee, Knoxville TRACE: Tennessee Research and Creative Exchange Doctoral Dissertations Graduate School 5-2011 Feature-based Image Comparison and Its Application in Wireless Visual Sensor Networks Yang Bai University of Tennessee Knoxville, Department of EECS, [email protected] Follow this and additional works at: https://trace.tennessee.edu/utk_graddiss Part of the Other Computer Engineering Commons, and the Robotics Commons Recommended Citation Bai, Yang, "Feature-based Image Comparison and Its Application in Wireless Visual Sensor Networks. " PhD diss., University of Tennessee, 2011. https://trace.tennessee.edu/utk_graddiss/946 This Dissertation is brought to you for free and open access by the Graduate School at TRACE: Tennessee Research and Creative Exchange. It has been accepted for inclusion in Doctoral Dissertations by an authorized administrator of TRACE: Tennessee Research and Creative Exchange. For more information, please contact [email protected]. To the Graduate Council: I am submitting herewith a dissertation written by Yang Bai entitled "Feature-based Image Comparison and Its Application in Wireless Visual Sensor Networks." I have examined the final electronic copy of this dissertation for form and content and recommend that it be accepted in partial fulfillment of the equirr ements for the degree of Doctor of Philosophy, with a major in Computer Engineering. Hairong Qi, Major Professor We have read this dissertation and recommend its acceptance: Mongi A. Abidi, Qing Cao, Steven Wise Accepted for the Council: Carolyn R. Hodges Vice Provost and Dean of the Graduate School (Original signatures are on file with official studentecor r ds.) Feature-based Image Comparison and Its Application in Wireless Visual Sensor Networks A Dissertation Presented for the Doctor of Philosophy Degree The University of Tennessee, Knoxville Yang Bai May 2011 Copyright c 2011 by Yang Bai All rights reserved.
    [Show full text]
  • Robust Edge Aware Descriptor for Image Matching
    Robust Edge Aware Descriptor for Image Matching Rouzbeh Maani, Sanjay Kalra, Yee-Hong Yang Department of Computing Science, University of Alberta, Edmonton, Canada Abstract. This paper presents a method called Robust Edge Aware De- scriptor (READ) to compute local gradient information. The proposed method measures the similarity of the underlying structure to an edge using the 1D Fourier transform on a set of points located on a circle around a pixel. It is shown that the magnitude and the phase of READ can well represent the magnitude and orientation of the local gradients and present robustness to imaging effects and artifacts. In addition, the proposed method can be efficiently implemented by kernels. Next, we de- fine a robust region descriptor for image matching using the READ gra- dient operator. The presented descriptor uses a novel approach to define support regions by rotation and anisotropical scaling of the original re- gions. The experimental results on the Oxford dataset and on additional datasets with more challenging imaging effects such as motion blur and non-uniform illumination changes show the superiority and robustness of the proposed descriptor to the state-of-the-art descriptors. 1 Introduction Local feature detection and description are among the most important tasks in computer vision. The extracted descriptors are used in a variety of applica- tions such as object recognition and image matching, motion tracking, facial expression recognition, and human action recognition. The whole process can be divided into two major tasks: region detection, and region description. The goal of the first task is to detect regions that are invariant to a class of transforma- tions such as rotation, change of scale, and change of viewpoint.
    [Show full text]
  • Computer Vision: Edge Detection
    Edge Detection Edge detection Convert a 2D image into a set of curves • Extracts salient features of the scene • More compact than pixels Origin of Edges surface normal discontinuity depth discontinuity surface color discontinuity illumination discontinuity Edges are caused by a variety of factors Edge detection How can you tell that a pixel is on an edge? Profiles of image intensity edges Edge detection 1. Detection of short linear edge segments (edgels) 2. Aggregation of edgels into extended edges (maybe parametric description) Edgel detection • Difference operators • Parametric-model matchers Edge is Where Change Occurs Change is measured by derivative in 1D Biggest change, derivative has maximum magnitude Or 2nd derivative is zero. Image gradient The gradient of an image: The gradient points in the direction of most rapid change in intensity The gradient direction is given by: • how does this relate to the direction of the edge? The edge strength is given by the gradient magnitude The discrete gradient How can we differentiate a digital image f[x,y]? • Option 1: reconstruct a continuous image, then take gradient • Option 2: take discrete derivative (finite difference) How would you implement this as a cross-correlation? The Sobel operator Better approximations of the derivatives exist • The Sobel operators below are very commonly used -1 0 1 1 2 1 -2 0 2 0 0 0 -1 0 1 -1 -2 -1 • The standard defn. of the Sobel operator omits the 1/8 term – doesn’t make a difference for edge detection – the 1/8 term is needed to get the right gradient
    [Show full text]
  • Features Extraction in Context Based Image Retrieval
    =================================================================== Engineering & Technology in India www.engineeringandtechnologyinindia.com Vol. 1:2 March 2016 =================================================================== FEATURES EXTRACTION IN CONTEXT BASED IMAGE RETRIEVAL A project report submitted by ARAVINDH A S J CHRISTY SAMUEL in partial fulfilment for the award of the degree of BACHELOR OF TECHNOLOGY in ELECTRONICS AND COMMUNICATION ENGINEERING under the supervision of Mrs. M. A. P. MANIMEKALAI, M.E. SCHOOL OF ELECTRICAL SCIENCES KARUNYA UNIVERSITY (Karunya Institute of Technology and Sciences) (Declared as Deemed to be University under Sec-3 of the UGC Act, 1956) Karunya Nagar, Coimbatore - 641 114, Tamilnadu, India APRIL 2015 Engineering & Technology in India www.engineeringandtechnologyinindia.com Vol. 1:2 March 2016 Aravindh A S. and J. Christy Samuel Features Extraction in Context Based Image Retrieval 3 BONA FIDE CERTIFICATE Certified that this project report “Features Extraction in Context Based Image Retrieval” is the bona fide work of ARAVINDH A S (UR11EC014), and CHRISTY SAMUEL J. (UR11EC026) who carried out the project work under my supervision during the academic year 2014-2015. Mrs. M. A. P. MANIMEKALAI, M.E SUPERVISOR Assistant Professor Department of ECE School of Electrical Sciences Dr. SHOBHA REKH, M.E., Ph.D., HEAD OF THE DEPARTMENT Professor Department of ECE School of Electrical Sciences Engineering & Technology in India www.engineeringandtechnologyinindia.com Vol. 1:2 March 2016 Aravindh A S. and J. Christy Samuel Features Extraction in Context Based Image Retrieval 4 ACKNOWLEDGEMENT First and foremost, we praise and thank ALMIGHTY GOD whose blessings have bestowed in us the will power and confidence to carry out our project. We are grateful to our most respected Founder (late) Dr.
    [Show full text]
  • Vision Review: Image Processing
    Vision Review: Image Processing Course web page: www.cis.udel.edu/~cer/arv September 17, 2002 Announcements • Homework and paper presentation guidelines are up on web page • Readings for next Tuesday: Chapters 6, 11.1, and 18 • For next Thursday: “Stochastic Road Shape Estimation” Computer Vision Review Outline • Image formation • Image processing • Motion & Estimation • Classification Outline •Images • Binary operators •Filtering – Smoothing – Edge, corner detection • Modeling, matching • Scale space Images • An image is a matrix of pixels Note: Matlab uses • Resolution – Digital cameras: 1600 X 1200 at a minimum – Video cameras: ~640 X 480 • Grayscale: generally 8 bits per pixel → Intensities in range [0…255] • RGB color: 3 8-bit color planes Image Conversion •RGB → Grayscale: Mean color value, or weight by perceptual importance (Matlab: rgb2gray) •Grayscale → Binary: Choose threshold based on histogram of image intensities (Matlab: imhist) Color Representation • RGB, HSV (hue, saturation, value), YUV, etc. • Luminance: Perceived intensity • Chrominance: Perceived color – HS(V), (Y)UV, etc. – Normalized RGB removes some illumination dependence: Binary Operations • Dilation, erosion (Matlab: imdilate, imerode) – Dilation: All 0’s next to a 1 → 1 (Enlarge foreground) – Erosion: All 1’s next to a 0 → 0 (Enlarge background) • Connected components – Uniquely label each n-connected region in binary image – 4- and 8-connectedness –Matlab: bwfill, bwselect • Moments: Region statistics – Zeroth-order: Size – First-order: Position (centroid)
    [Show full text]