Edge Detection Low Level

Total Page:16

File Type:pdf, Size:1020Kb

Edge Detection Low Level Image Processing - Lesson 10 Image Processing - Computer Vision Edge Detection Low Level • Edge detection masks Image Processing representation, compression,transmission • Gradient Detectors • Compass Detectors image enhancement • Second Derivative - Laplace detectors • Edge Linking edge/feature finding • Hough Transform image "understanding" Computer Vision High Level UFO - Unidentified Flying Object Point Detection -1 -1 -1 Convolution with: -1 8 -1 -1 -1 -1 Large Positive values = light point on dark surround Large Negative values = dark point on light surround Example: 5 5 5 5 5 -1 -1 -1 5 5 5 100 5 * -1 8 -1 5 5 5 5 5 -1 -1 -1 0 0 -95 -95 -95 = 0 0 -95 760 -95 0 0 -95 -95 -95 Edge Definition Edge Detection Line Edge Line Edge Detectors -1 -1 -1 -1 -1 2 -1 2 -1 2 -1 -1 2 2 2 -1 2 -1 -1 2 -1 -1 2 -1 -1 -1 -1 2 -1 -1 -1 2 -1 -1 -1 2 gray value gray x edge edge Step Edge Step Edge Detectors -1 1 -1 -1 1 1 -1 -1 -1 -1 -1 -1 -1 -1 -1 1 -1 -1 1 1 -1 -1 -1 -1 1 1 1 1 -1 1 -1 -1 1 1 1 1 1 1 gray value gray -1 -1 1 1 1 -1 1 1 x edge edge Example Edge Detection by Differentiation Step Edge detection by differentiation: 1D image f(x) gray value gray x 1st derivative f'(x) threshold |f'(x)| - threshold Pixels that passed the threshold are -1 -1 1 1 -1 -1 -1 -1 Edge Pixels -1 -1 1 1 -1 -1 -1 -1 -1 -1 1 1 1 1 1 1 -1 -1 1 1 1 -1 1 1 Gradient Edge Detection Gradient Edge - Examples ∂f ∂x ∇ f((x,y) = Gradient ∂f ∂y ∂f ∇ f((x,y) = ∂x , 0 f 2 f 2 Gradient Magnitude ∂ + ∂ √( ∂ x ) ( ∂ y ) ∂f ∂f Gradient Direction tg-1 ( / ) ∂y ∂x ∂f ∇ f((x,y) = 0 , ∂y Differentiation in Digital Images Example Edge horizontal - differentiation approximation: ∂f(x,y) Original FA = ∂x = f(x,y) - f(x-1,y) convolution with [ 1 -1 ] Gradient-X Gradient-Y vertical - differentiation approximation: ∂f(x,y) FB = ∂y = f(x,y) - f(x,y-1) convolution with 1 -1 Gradient-Magnitude Gradient-Direction Gradient (FA , FB) 2 2 1/2 Magnitude ((FA ) + (FB) ) Approx. Magnitude |FA | + |FB| Roberts Edge Detector Sobel Edge Detector F = f(x,y) - f(x-1,y-1) A -1 0 1 -1 -2 -1 FB = f(x-1,y) - f(x,y-1) A = -2 0 2 B = 0 0 0 -1 0 1 1 2 1 1 0 0 1 A = B = 0 -1 -1 0 Roberts and other 2x2 operators are sensitive to noise. Isotropic Edge Detector Prewitt Edge Detector -1 0 1 -1 -√2 -1 A = -√2 0 √2 B = 0 0 0 -1 0 1 -1 -1 -1 -1 0 1 1 √2 1 A = -1 0 1 B = 0 0 0 -1 0 1 1 1 1 Smoothed operators Edge Detector - Comparison Compass Operators 1 1 1 1 1 0 1 0 -1 0 -1 -1 Original 0 0 0 1 0 -1 1 0 -1 1 0 -1 -1 -1 -1 0 -1 -1 1 0 -1 1 1 0 N NW WSW -1 -1 -1 -1 -1 0 -1 0 1 0 1 1 Roberts 0 0 0 -1 0 1 -1 0 1 -1 0 1 1 1 1 0 1 1 -1 0 1 -1 -1 0 S SE ENE Given k operators, g (x,y) is the image obtained by Sobel k convolving f(x,y) with the kth operator. The gradient is defined as: g(x,y) = max gk(x,y) Prewitt k k defines the edge direction Magnitude Negative Magnitude Various Compass Operators: Edge Detector - Comparison 1 1 1 1 1 1 1 2 1 Original 1 -2 1 0 0 0 0 0 0 -1 -1 -1 -1 -1 -1 -1 -2 -1 5 5 5 Sobel -3 0 -3 Kirsch Edge Detector -3 -3 -3 Kirsch (Compass) Direction Map Magnitude Negative Magnitude Laplacian Operators Derivatives Approximation of second derivative (horizontal): ∂f2(x,y) ∂2x = f''(x,y) = f'(x+1,y) - f'(x,y) = = [f(x+1,y) - f(x,y)] - [f(x,y) - f(x-1,y)] f(x) = f(x+1,y) - 2 f(x,y) + f(x-1,y) convolution with: [ 1 -2 1] Approximation of second derivative (vertical): f (x) convolution with: 1 -2 1 2 2 f (x) 2 ∂ ∂ Laplacian Operator ∇ = ( ∂ x 2 + ∂ y 2 ) 0 -1 0 zero crossing convolution with: -1 4 -1 0 -1 0 Variations on Laplace Operators: Example of Laplacian Edge Detection 1 -2 1 0 -1 0 -1 -1 -1 -2 4 -2 -1 4 -1 -1 8 -1 1 -2 1 0 -1 0 -1 -1 -1 All are approximations of: Laplacian ~ Difference of gaussians Laplacian Operator Laplacian Filter (Image Domain) (Frequency Domain) - = FFT DOG = Difference of Gaussians - = FFT - = Enhancement Using the Laplacian Operator Examples of enhanced images Original Laplacian Image Edge Image f(x) ∂f 1st derivative ∂x 2nd derivative ∂2f (Laplacian) ∂x2 2 - ∂ f ∂x2 Scaled Laplacian Image Enhanced Image 2 f(x) - ∂ f ∂x2 Mach Bands Edge Detection - Noise Issues Edge Detection - Noise Issues f f ( x ) g d f ( x ) g ∗ f dx ∂ ()g ∗ f ∂x ∂ Peaks in () g ∗ f mark the edge ∂x Edge Detection - Noise Issues Edge Detection - Noise Issues From the convolution theorem we have: ∂ ⎛ ∂ ⎞ ()h ∗ f = ⎜ h ⎟ ∗ f ∂x ⎝ ∂x ⎠ f ∂ 2 g f ∂x 2 ⎛ ∂ 2 ⎞ ∂ ⎜ g ⎟ ∗ f h ⎜ 2 ⎟ ∂x ⎝ ∂x ⎠ ⎛ ∂ ⎞ ⎜ h ⎟ ∗ f ⎝ ∂x ⎠ ∂ 2 Zero Crossings in () g ∗ f mark the edge ∂x 2 Canny Edge Detector Canny Edge Detector - Step 1 1) Convolve image with derivative of a Gaussian Convolve image with derivative of a Gaussian 2) Perform NonMaximum suppression. 3) Perform Hysteresis Thresholding. f(x,y) * G'(x,y) ≅ f(x,y) * G'x(x,y) + f(x,y) * G'y(x,y) Fa = f(x,y) * Gx(x,y) * Dx(x,y) = f(x,y) *1 2 4 2 1 * 1 -1 Fb = f(x,y) * Gy(x,y) * Dy(x,y) 1 2 1 = f(x,y) *4 * -1 2 1 2 2 1/2 Magnitude ((FA ) + (FB) ) Canny Edge Detector - Step 1 Canny Edge Detector - Step 2 NonMaximum suppression. Original Remove edges that are NOT local Maxima in the gradient direction. x x x x x x f * Gx f * Gy Local Maxima Local Maxima Fa = Fb = f*Gx*Dx f*Gy*Dy Magnitude 2 2 1/2 ((FA ) + (FB) ) Canny Edge Detector - Step 2 Canny Edge Detector - Step 3 Hysteresis Thresholding. NonMaximum Suppression. Magnitude NonMaximum Supression Edge Pixel High Threshold Possible Edge Pixel Low Threshold Non-Edge Pixel Include possible edge pixels that are adjacent to edge pixels. Canny Edge Detector - Step 3 Effect of Scale Effect of σ (Gaussian kernel size) Magnitude (NonMax original Suppressed) Low Threshold High Threshold Canny with σ = 1 Canny with σ = 2 Hysteresis Threshold Scale Space Edge Completion Witkin 83 first derivative peaks Gaussian filtered signal Properties of Scale Space: • Position of edges may change with scale. • Edges may merge with increase in scale. • Edges do NOT split with increase in scale. Edge Linking Examples of edge linking (x,y) is an edge pixel. Search for neighboring edge pixels that are "similar". Original Sobel Vertical Similarity: Similarity in Edge Orientation Similarity in Edge strength (Gradient Amplitude) Perform Edge Following along similar edge pixels. (as in Contour Following in binary images). Sobel Horizontal Linked Edges Edge Points linked: Gradient Value > 25 Gradient direction within 15% Hough Transform Problem: Edges are not lines even when linked. Image Domain Hough Domain s y (θ, s) s θ x θ straight line Hough Transform Edge pixels are not ellipses even when linked. y = ax+b s = xcos(θ)+ysin(θ) Hough Transform Hough Transform Image Domain Hough Domain y = ax+b s y (θ, s) y s s θ x θ x θ straight line Hough Transform many points on a line = many lines in the Hough transform space which intersect at 1 point. s y x θ single point = many possible lines Hough Transform Example Hough Transform Example Original Edges Original square image Hough Transform Hough Transform s (s,θ) space θ Results1 Results2 Results3 Reconstructed line segments Hough Transform Example Hough Transform for Circles Image Domain Hough Domain Original (x0,y0) r (x0,y0,r) y r y0 x x0 circle Hough Transform Edges 2 2 2 r = (x-x0) +(y-y0) Result Nice demo: http://www.markschulze.net/java/hough/index.html 3D Perception - Depth Perception Impossible Figures (Escher).
Recommended publications
  • Edge Detection of an Image Based on Extended Difference of Gaussian
    American Journal of Computer Science and Technology 2019; 2(3): 35-47 http://www.sciencepublishinggroup.com/j/ajcst doi: 10.11648/j.ajcst.20190203.11 ISSN: 2640-0111 (Print); ISSN: 2640-012X (Online) Edge Detection of an Image Based on Extended Difference of Gaussian Hameda Abd El-Fattah El-Sennary1, *, Mohamed Eid Hussien1, Abd El-Mgeid Amin Ali2 1Faculty of Science, Aswan University, Aswan, Egypt 2Faculty of Computers and Information, Minia University, Minia, Egypt Email address: *Corresponding author To cite this article: Hameda Abd El-Fattah El-Sennary, Abd El-Mgeid Amin Ali, Mohamed Eid Hussien. Edge Detection of an Image Based on Extended Difference of Gaussian. American Journal of Computer Science and Technology. Vol. 2, No. 3, 2019, pp. 35-47. doi: 10.11648/j.ajcst.20190203.11 Received: November 24, 2019; Accepted: December 9, 2019; Published: December 20, 2019 Abstract: Edge detection includes a variety of mathematical methods that aim at identifying points in a digital image at which the image brightness changes sharply or, more formally, has discontinuities. The points at which image brightness changes sharply are typically organized into a set of curved line segments termed edges. The same problem of finding discontinuities in one-dimensional signals is known as step detection and the problem of finding signal discontinuities over time is known as change detection. Edge detection is a fundamental tool in image processing, machine vision and computer vision, particularly in the areas of feature detection and feature extraction. It's also the most important parts of image processing, especially in determining the image quality.
    [Show full text]
  • Fast Image Registration Using Pyramid Edge Images
    Fast Image Registration Using Pyramid Edge Images Kee-Baek Kim, Jong-Su Kim, Sangkeun Lee, and Jong-Soo Choi* The Graduate School of Advanced Imaging Science, Multimedia and Film, Chung-Ang University, Seoul, Korea(R.O.K) *Corresponding author’s E-mail address: [email protected] Abstract: Image registration has been widely used in many image processing-related works. However, the exiting researches have some problems: first, it is difficult to find the accurate information such as translation, rotation, and scaling factors between images; and it requires high computational cost. To resolve these problems, this work proposes a Fourier-based image registration using pyramid edge images and a simple line fitting. Main advantages of the proposed approach are that it can compute the accurate information at sub-pixel precision and can be carried out fast for image registration. Experimental results show that the proposed scheme is more efficient than other baseline algorithms particularly for the large images such as aerial images. Therefore, the proposed algorithm can be used as a useful tool for image registration that requires high efficiency in many fields including GIS, MRI, CT, image mosaicing, and weather forecasting. Keywords: Image registration; FFT; Pyramid image; Aerial image; Canny operation; Image mosaicing. registration information as image size increases [3]. 1. Introduction To solve these problems, this work employs a pyramid-based image decomposition scheme. Image registration is a basic task in image Specifically, first, we use the Gaussian filter to processing to combine two or more images that are remove the noise because it causes many undesired partially overlapped.
    [Show full text]
  • Scale Invariant Feature Transform - Scholarpedia 2015-04-21 15:04
    Scale Invariant Feature Transform - Scholarpedia 2015-04-21 15:04 Scale Invariant Feature Transform +11 Recommend this on Google Tony Lindeberg (2012), Scholarpedia, 7(5):10491. doi:10.4249/scholarpedia.10491 revision #149777 [link to/cite this article] Prof. Tony Lindeberg, KTH Royal Institute of Technology, Stockholm, Sweden Scale Invariant Feature Transform (SIFT) is an image descriptor for image-based matching and recognition developed by David Lowe (1999, 2004). This descriptor as well as related image descriptors are used for a large number of purposes in computer vision related to point matching between different views of a 3-D scene and view-based object recognition. The SIFT descriptor is invariant to translations, rotations and scaling transformations in the image domain and robust to moderate perspective transformations and illumination variations. Experimentally, the SIFT descriptor has been proven to be very useful in practice for image matching and object recognition under real-world conditions. In its original formulation, the SIFT descriptor comprised a method for detecting interest points from a grey- level image at which statistics of local gradient directions of image intensities were accumulated to give a summarizing description of the local image structures in a local neighbourhood around each interest point, with the intention that this descriptor should be used for matching corresponding interest points between different images. Later, the SIFT descriptor has also been applied at dense grids (dense SIFT) which have been shown to lead to better performance for tasks such as object categorization, texture classification, image alignment and biometrics . The SIFT descriptor has also been extended from grey-level to colour images and from 2-D spatial images to 2+1-D spatio-temporal video.
    [Show full text]
  • Sobel Operator and Canny Edge Detector ECE 480 Fall 2013 Team 4 Daniel Kim
    P a g e | 1 Sobel Operator and Canny Edge Detector ECE 480 Fall 2013 Team 4 Daniel Kim Executive Summary In digital image processing (DIP), edge detection is an important subject matter. There are numerous edge detection methods such as Prewitt, Kirsch, and Robert cross. Two different types of edge detectors are explored and analyzed in this paper: Sobel operator and Canny edge detector. The performance of two edge detection methods are then compared with several input images. Keywords : Digital Image Processing, Edge Detection, Sobel Operator, Canny Edge Detector, Computer Vision, OpenCV P a g e | 2 Table of Contents 1| Introduction……………………………………………………………………………. 3 2|Sobel Operator 2.1 Background……………………………………………………………………..3 2.2 Example………………………………………………………………………...4 3|Canny Edge Detector 3.1 Background…………………………………………………………………….5 3.2 Example………………………………………………………………………...5 4|Comparison of Sobel and Canny 4.1 Discussion………………………………………………………………………6 4.2 Example………………………………………………………………………...7 5|Conclusion………………………………………………………………………............7 6|Appendix 6.1 OpenCV Sobel tutorial code…............................................................................8 6.2 OpenCV Canny tutorial code…………..…………………………....................9 7|Reference………………………………………………………………………............10 P a g e | 3 1) Introduction Edge detection is a crucial step in object recognition. It is a process of finding sharp discontinuities in an image. The discontinuities are abrupt changes in pixel intensity which characterize boundaries of objects in a scene. In short, the goal of edge detection is to produce a line drawing of the input image. The extracted features are then used by computer vision algorithms, e.g. recognition and tracking. A classical method of edge detection involves the use of operators, a two dimensional filter. An edge in an image occurs when the gradient is greatest.
    [Show full text]
  • Fast Almost-Gaussian Filtering
    Fast Almost-Gaussian Filtering Peter Kovesi Centre for Exploration Targeting School of Earth and Environment The University of Western Australia 35 Stirling Highway Crawley WA 6009 Email: [email protected] Abstract—Image averaging can be performed very efficiently paper describes how the fixed low cost of averaging achieved using either separable moving average filters or by using summed through separable moving average filters, or via summed area area tables, also known as integral images. Both these methods tables, can be exploited to achieve a good approximation allow averaging to be performed at a small fixed cost per pixel, independent of the averaging filter size. Repeated filtering with to Gaussian filtering also at a small fixed cost per pixel, averaging filters can be used to approximate Gaussian filtering. independent of filter size. Thus a good approximation to Gaussian filtering can be achieved Summed area tables were devised by Crow in 1984 [12] but at a fixed cost per pixel independent of filter size. This paper only recently introduced to the computer vision community by describes how to determine the averaging filters that one needs Viola and Jones [13]. A summed area table, or integral image, to approximate a Gaussian with a specified standard deviation. The design of bandpass filters from the difference of Gaussians can be generated by computing the cumulative sums along the is also analysed. It is shown that difference of Gaussian bandpass rows of an image and then computing the cumulative sums filters share some of the attributes of log-Gabor filters in that down the columns.
    [Show full text]
  • Design of Improved Canny Edge Detection Algorithm
    IJournals: International Journal of Software & Hardware Research in Engineering ISSN-2347-4890 Volume 3 Issue 8 August, 2015 Design of Improved Canny Edge Detection Algorithm Deepa Krushnappa Maladakara; H R Vanamala M.Tech 4th SEM Student; Associate Professor PESIT Bengaluru; PESIT Bengaluru [email protected]; [email protected] ABSTRACT— An improved Canny edge detection detector, Laplacian of Gaussian, etc. Among these, algorithm is implemented to detect the edges with Canny edge detection algorithm is most extensively complexity reduction and without any performance used in industries, as it is highly performance oriented degradation. The existing algorithms for edge detection and provides low error bit rate. are Sobel operator, Robert Operator, Prewitt Operator, The Canny edge detection operator was developed by Laplacian of Gaussian Operator, have various draw John F. Canny in 1986, an approach to derive an backs in edge detection methodologies with respect to optimal edge detector to deal with step edges corrupted noise, performance, etc. Original Canny edge detection by a white Gaussian noise. algorithm based on frame level statistics overcomes the An optimal edge detector should have the flowing 1 drawbacks giving high accuracy but is computationally criteria . more intensive to implement in hardware. Hence a 1) Good Detection: Must have a minimum error novel method is suggested to decrease the complexity detection rate, i.e. there are no significant edges to be without any performance degradation compared to the missed and no false edges are detected. original algorithm. Further enhancements to the 2) Good Localization: The space between actual and proposed algorithm can be made by parallel processing located edges should be minimal.
    [Show full text]
  • Edge Detection
    Edge Detection Outline Part 1 • Introduction Part 2 • Local Edge Operators • Marr-Hildreth Edge Detector – Gradient of image intensity • Multiscale Processing – Robert, Sobel and Prewitt operators – Second Derivative Operators • Canny Edge Detector – Laplacian of Gaussian • Edge Detection Performance References: D. Huttenlocher, “Edge Detection” (http://www.cs.wisc.edu/~cs766-1/readings/edge-detection- huttenlocher.ps) R. Gonzalez, R. Woods, “Digital Image Processing”, Addison Wesley, 1993. D. Marr, E. Hildreth, “Theory of Edge detection”, Proc. R. Soc. Lond., B. 207, 187-217, 1980. Image Processing Applications Edge Detection (by A.Campilho) 1 Edge Detection Introduction Initial considerations Edges are significant local intensity changes in the image and are important features to analyse an image. They are important clues to separate regions within an object or to identify changes in illumination or colour. They are an important feature in the early vision stages in the human eye. There are many different geometric and optical physical properties in the scene that can originate an edge in an image. Geometric: • Object boundary: eg a discontinuity in depth and/or surface colour or texture • Surface boundary: eg a discontinuity in surface orientation and/or colour or texture • Occlusion boundary: eg a discontinuity in depth and/or surface colour or texture Optical: • Specularity: eg direct reflection of light, such as a polished metallic surface • Shadows: from other objects or part of the same object • Interreflections: from other objects or part sof the same object • Surface markings, texture or colour changes Image Processing Applications Edge Detection (by A.Campilho) 2 Edge Detection Introduction Goals of edge detection: Primary To extract information about the two-dimensional projection of a 3D scene.
    [Show full text]
  • Edges and Binary Image Analysis
    1/25/2017 Edges and Binary Image Analysis Thurs Jan 26 Kristen Grauman UT Austin Today • Edge detection and matching – process the image gradient to find curves/contours – comparing contours • Binary image analysis – blobs and regions 1 1/25/2017 Gradients -> edges Primary edge detection steps: 1. Smoothing: suppress noise 2. Edge enhancement: filter for contrast 3. Edge localization Determine which local maxima from filter output are actually edges vs. noise • Threshold, Thin Kristen Grauman, UT-Austin Thresholding • Choose a threshold value t • Set any pixels less than t to zero (off) • Set any pixels greater than or equal to t to one (on) 2 1/25/2017 Original image Gradient magnitude image 3 1/25/2017 Thresholding gradient with a lower threshold Thresholding gradient with a higher threshold 4 1/25/2017 Canny edge detector • Filter image with derivative of Gaussian • Find magnitude and orientation of gradient • Non-maximum suppression: – Thin wide “ridges” down to single pixel width • Linking and thresholding (hysteresis): – Define two thresholds: low and high – Use the high threshold to start edge curves and the low threshold to continue them • MATLAB: edge(image, ‘canny’); • >>help edge Source: D. Lowe, L. Fei-Fei The Canny edge detector original image (Lena) Slide credit: Steve Seitz 5 1/25/2017 The Canny edge detector norm of the gradient The Canny edge detector thresholding 6 1/25/2017 The Canny edge detector How to turn these thick regions of the gradient into curves? thresholding Non-maximum suppression Check if pixel is local maximum along gradient direction, select single max across width of the edge • requires checking interpolated pixels p and r 7 1/25/2017 The Canny edge detector Problem: pixels along this edge didn’t survive the thresholding thinning (non-maximum suppression) Credit: James Hays 8 1/25/2017 Hysteresis thresholding • Use a high threshold to start edge curves, and a low threshold to continue them.
    [Show full text]
  • Lecture 10 Detectors and Descriptors
    This lecture is about detectors and descriptors, which are the basic building blocks for many tasks in 3D vision and recognition. We’ll discuss Lecture 10 some of the properties of detectors and descriptors and walk through examples. Detectors and descriptors • Properties of detectors • Edge detectors • Harris • DoG • Properties of descriptors • SIFT • HOG • Shape context Silvio Savarese Lecture 10 - 16-Feb-15 Previous lectures have been dedicated to characterizing the mapping between 2D images and the 3D world. Now, we’re going to put more focus From the 3D to 2D & vice versa on inferring the visual content in images. P = [x,y,z] p = [x,y] 3D world •Let’s now focus on 2D Image The question we will be asking in this lecture is - how do we represent images? There are many basic ways to do this. We can just characterize How to represent images? them as a collection of pixels with their intensity values. Another, more practical, option is to describe them as a collection of components or features which correspond to “interesting” regions in the image such as corners, edges, and blobs. Each of these regions is characterized by a descriptor which captures the local distribution of certain photometric properties such as intensities, gradients, etc. Feature extraction and description is the first step in many recent vision algorithms. They can be considered as building blocks in many scenarios The big picture… where it is critical to: 1) Fit or estimate a model that describes an image or a portion of it 2) Match or index images 3) Detect objects or actions form images Feature e.g.
    [Show full text]
  • Edge Detection 1-D &
    Edge Detection CS485/685 Computer Vision Dr. George Bebis Modified by Guido Gerig for CS/BIOEN 6640 U of Utah www.cse.unr.edu/~bebis/CS485/Lectures/ EdgeDetection.ppt Edge detection is part of segmentation image human segmentation gradient magnitude • Berkeley segmentation database: http://www.eecs.berkeley.edu/Research/Projects/CS/vision/grouping/segbench/ Goal of Edge Detection • Produce a line “drawing” of a scene from an image of that scene. Why is Edge Detection Useful? • Important features can be extracted from the edges of an image (e.g., corners, lines, curves). • These features are used by higher-level computer vision algorithms (e.g., recognition). Modeling Intensity Changes • Step edge: the image intensity abruptly changes from one value on one side of the discontinuity to a different value on the opposite side. Modeling Intensity Changes (cont’d) • Ramp edge: a step edge where the intensity change is not instantaneous but occur over a finite distance. Modeling Intensity Changes (cont’d) • Ridge edge: the image intensity abruptly changes value but then returns to the starting value within some short distance (i.e., usually generated by lines). Modeling Intensity Changes (cont’d) • Roof edge: a ridge edge where the intensity change is not instantaneous but occur over a finite distance (i.e., usually generated by the intersection of two surfaces). Edge Detection Using Derivatives • Often, points that lie on an edge are detected by: (1) Detecting the local maxima or minima of the first derivative. 1st derivative (2) Detecting the zero-crossings of the second derivative. 2nd derivative Image Derivatives • How can we differentiate a digital image? – Option 1: reconstruct a continuous image, f(x,y), then compute the derivative.
    [Show full text]
  • Anatomy of the SIFT Method Ives Rey Otero
    Anatomy of the SIFT method Ives Rey Otero To cite this version: Ives Rey Otero. Anatomy of the SIFT method. General Mathematics [math.GM]. École normale supérieure de Cachan - ENS Cachan, 2015. English. NNT : 2015DENS0044. tel-01226489 HAL Id: tel-01226489 https://tel.archives-ouvertes.fr/tel-01226489 Submitted on 9 Nov 2015 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Ecole´ Normale Sup´erieure de Cachan, France Anatomy of the SIFT method A dissertation presented by Ives Rey-Otero in fulfillment of the requirements for the degree of Doctor of Philosophy in the subject of Applied Mathematics Committee in charge Referees Coloma Ballester - Universitat Pompeu Fabra, ES Pablo Muse´ - Universidad de la Rep´ublica, UY Joachim Weickert - Universit¨at des Saarlandes, DE Advisors Jean-Michel Morel - ENS de Cachan, FR Mauricio Delbracio - Duke University, USA Examiners Patrick Perez´ - Technicolor Research, FR Fr´ed´eric Sur - Loria, FR ENSC-2015 603 September 2015 Version of November 5, 2015 at 12:27. Abstract of the Dissertation Ives Rey-Otero Anatomy of the SIFT method Under the direction of: Jean-Michel Morel and Mauricio Delbracio This dissertation contributes to an in-depth analysis of the SIFT method.
    [Show full text]
  • Edge Detection (Trucco, Chapt 4 and Jain Et Al., Chapt 5)
    Edge detection (Trucco, Chapt 4 AND Jain et al., Chapt 5) • Definition of edges -Edges are significant local changes of intensity in an image. -Edges typically occur on the boundary between twodifferent regions in an image. • Goal of edge detection -Produce a line drawing of a scene from an image of that scene. -Important features can be extracted from the edges of an image (e.g., corners, lines, curves). -These features are used by higher-levelcomputer vision algorithms (e.g., recogni- tion). -2- • What causes intensity changes? -Various physical events cause intensity changes. -Geometric events *object boundary (discontinuity in depth and/or surface color and texture) *surface boundary (discontinuity in surface orientation and/or surface color and texture) -Non-geometric events *specularity (direct reflection of light, such as a mirror) *shadows (from other objects or from the same object) *inter-reflections • Edge descriptors Edge normal: unit vector in the direction of maximum intensity change. Edge direction: unit vector to perpendicular to the edge normal. Edge position or center: the image position at which the edge is located. Edge strength: related to the local image contrast along the normal. -3- • Modeling intensity changes -Edges can be modeled according to their intensity profiles. Step edge: the image intensity abruptly changes from one value to one side of the discontinuity to a different value on the opposite side. Ramp edge: astep edge where the intensity change is not instantaneous but occur overafinite distance. Ridge edge: the image intensity abruptly changes value but then returns to the starting value within some short distance (generated usually by lines).
    [Show full text]