Image Denoising Using Non Linear Diffusion Tensors

Total Page:16

File Type:pdf, Size:1020Kb

Image Denoising Using Non Linear Diffusion Tensors 2011 8th International Multi-Conference on Systems, Signals & Devices IMAGE DENOISING USING NON LINEAR DIFFUSION TENSORS Benzarti Faouzi, Hamid Amiri Signal, Image Processing and Pattern Recognition: TSIRF (ENIT) Email: [email protected] , [email protected] ABSTRACT linear models. Many approaches have been proposed to remove the noise effectively while preserving the original Image denoising is an important pre-processing step for image details and features as much as possible. In the past many image analysis and computer vision system. It few years, the use of non linear PDEs methods involving refers to the task of recovering a good estimate of the true anisotropic diffusion has significantly grown and image from a degraded observation without altering and becomes an important tool in contemporary image changing useful structure in the image such as processing. The key idea behind the anisotropic diffusion discontinuities and edges. In this paper, we propose a new is to incorporate an adaptative smoothness constraint in approach for image denoising based on the combination the denoising process. That is, the smooth is encouraged of two non linear diffusion tensors. One allows diffusion in a homogeneous region and discourage across along the orientation of greatest coherences, while the boundaries, in order to preserve the natural edge of the other allows diffusion along orthogonal directions. The image. One of the most successful tools for image idea is to track perfectly the local geometry of the denoising is the Total Variation (TV) model [9] [8][10] degraded image and applying anisotropic diffusion and the anisotropic smoothing model [1] which has since mainly along the preferred structure direction. To been expanded and improved upon [3] [5][20]. Over the illustrate the effective performance of our model, we years, other very interesting denoising methods have been present some experimental results on a test and real emerged such as: Bilateral filter and its derivatives photographic color images. [6][11][12]. In our work, we address image denoising problem by using the so-called structure tensors [14] Index Terms— Image denoising, Non-linear which have proven their effectiveness in several areas diffusion, Structure tensors, PDEs. such as: texture segmentation [17], motion analysis [18] and corner detection [16][19][21]. The structure tensor provides a more powerful description of local pattern images better than a simple gradient. Based on its 1. INTRODUCTION eigenvalues and the corresponding eigenvectors, the tensor summarizes the predominant directions of the Image denoising has been one of the most important and gradient in a specified neighborhood of a point and the widely studied problems in image processing and degree to which those directions are coherent. Our computer vision. The need to have a very good image contribution in this work lies in the use of two coupled quality is increasingly required with the advent of the diffusion tensors which allow a complete and coherence new technologies in a various areas such as multimedia, regularization process which significantly improves the medical image analysis, aerospace, video systems and quality image. others. Indeed, the acquired image is often marred by This paper is organized as follows. In Section 2, we noise which may have a multiple origins such as: thermal introduce the non linear diffusion PDEs and discuss the fluctuations; quantify effects and properties of various options that have been implemented for the communication channels. It affects the perceptual quality anisotropic diffusion. In Section 3, we present the non of the image, decreasing not only the appreciation of the linear structure tensor formalism and its mathematical image but also the performance of the task for which the concept. Section 4 focuses on our proposed denoising image has been intended. The challenge is to design approach. Numerical experiments and results on test and methods, which can selectively smooth a degraded image real photographic images are shown in Section 5. without altering edges, losing significant features and producing reliable results. Traditionally, linear models 2. NON LINEAR DIFFUSION PDE: OVERVIEW have been commonly used to reduce noise. It is shown that these methods perform well in the flat regions of In this section, we review some basic mathematical images but do not preserve edges and discontinuities concepts of the nonlinear diffusion PDEs proposed by which are often smeared out. In contrast, nonlinear Perona and Malik [1]. Let be the grayscaled intensity image with a diffusion time t, for the models can handle edges in a much better way than those ͩ ʚͬ, ͭ, ͨʛ: Ŵ ͌ 978-1-4577-0411-6/11/$26.00 ©2011 IEEE image domain . The nonlinear PDE equation is An extension of the nonlinear diffusion filtering to given by: ͦ vector-valued image (e.g. color image) has been proposed Ǩ ͌ [20][3]. It evolves: under the diffusion equations: ) x (1) Ŵ ͌ (3) "/ͩ Ɣ ͪ͘͝ ʚ͛ʚ|ǧͩ|ʛǧͩʛ ͣ͢(e.g. Initialʚ0, ∞ʛCondition) ) ͦ ͩʚͬ, ͭ, 0ʛ Ɣ ͩͤ ͣ͢ (e.g. reflecting boundary) "/ͩ$ Ɣ ͪ͘͝ ʚ ͛ʚ∑&Ͱͥ |ǧͩ&| ʛǧͩ$ʛ ͝ Ɣ 1. ͢ Where :denotes the i th component channels of "ͩ) Ɣ 0 ͣ͢ " x ʚ0, ∞ʛ where denotes the first derivative regarding the Note thatͩ$ , represents the luminanceͩ function, which) coupled ͦall vectors channels taking the diffusion"/ timeͩ: t; : denotes the gradient modulus and ∑&Ͱͥ |ǧͩ&| g( ) is a non-increasing function, known as the diffusivity strong correlations among channels . However, this |ǧͩ| function which allow isotropic diffusion in flat regions luminance function is not being able to detect iso- and no diffusion near edges. luminance contours. By developing the divergence term of (1), we obtain: 3. NON LINEAR DIFFUSION TENSOR ɒ (2) ɑɑ " ʚ|O0 |ʛ The non linear diffusion PDE seen previously does not "/ͩ Ɣ ͛ ʚ|ͩ |ʛͩVV ƍ | O0 | ͩ ]] give reliable information in the presence of flow-like Ɣ ͗VͩVV ƍ ͗]ͩ ]] structures (e.g. fingerprints). It would be desirable to where: and are respectively rotate the flow towards the orientation of interesting ʹ ʹ features. This can be easily achieved by using the the secondͩVV spatialƔ ͂ derivatives ͩ ]] ofƔ in͂ the directions of the gradient , and its orthogonal ; H structure tensor, also referred to the second moment ͩ denotes the Hessian of u. According to these definʹitions, matrix. For a multivalued image, the structure tensor has Ɣ ǧͩ/|ǧͩ| Ɣ on the image discontinuities, we have the diffusion along the following form: (normal to the edge) weighted with and Μ ͦ Μ (4) ɑɑ ΗͰͥ ΗΦa ΗͰͥ ΗΦa ΗΧa a diffusion along (tangential to theV edge) weighted ) ͳ ∑ u ∑ u u ͗ Ɣ ͛ ʚ|ͩ |ʛ ͍a ƔƳ∑$Ͱͥ ǧͩ$a ǧͩ$a Ʒ Ɣ ʨ Μ Μ ͦ ʩ with . To understand the principle of ∑ΗͰͥ uΗΦauΗΧa ∑ΗͰͥ uΗΧa ɑ With : the smoothed the anisotropic ͗] Ɣ ͛ ʚ|ͩ |ʛdiffusion,/|ǧͩ| let represent a contour C (figure 1) separating two homogeneous regions of the version ǧͩof$a Ɣthe ͅ a dzgradient ǧͩ$ Ɣ ͅawhichdz ʚuix , uisiy ʛ obtained by a image, the isophote lines (e.g. level curves of equal gray- convolution with a Gaussian kernel . The structure levels) correspond to u(x,y) = c. In this case, the vector scale determines the size of the resultingͅa flow-like patterns. Increasing gives an increased distance is normal to the contour C, the set is then a moving orthonormal basis whose configuration depends on the between the resulting flow lines. ʚ, ʛ current coordinate point (x, y). In the neighborhood of a These new gradient features allow a more precise contour C, the image presents a strong gradient. To better description of the local gradient characteristics. However, preserve these discontinuities, it is preferable to diffuse it is more convenient to use a smoothed version of , only in the direction parallel to C (i.e. in the -direction). that is: ͍a In this case, we have to inhibit the coefficient of (5) ͥͥ͞ ͥͦ͞ (e.g. , and to suppose that the coefficient of VV ̈́` Ɣ ͅ` dz ͍a ƔƬ ư ͩ ͦͥ͞ ͦͦ͞ does not͗V Ɣ vanish. 0ʛ ͩ]] Where : a Gaussian kernel with standard deviation . The integrationͅ` scale averages orientation information. Therefore, it helps to stabilize the directional behavior of u(x,y)>c η the filter. In particular, it is possible to close interrupted (x o,y o) lines if is equal or larger than the gap size. In order to (x 1 , y1) ξ enhance coherent structures, the integration scale u(x,y)<c should be larger than the structure scale ( u(x,y)=c .In summary, the convolution with the Gaussian ͙ͬ: Ɣ 3ʛkernels , make the structure tensor measure more Figure 1 Image contour and its moving orthonormal basis coherent. ͅ aTo, ͅ `go further into the formalism, the structure tensor can be written over its eigenvalues ( , ) ʚ, ʛ So it appears the following conditions for the choice of and eigenvectors̈́` , that is: ͮ ͯ the g( ) functions to be edge preserving : ʚͮ, ͯʛ : avoids inverse diffusion λ θ ɑɑ ɑ θ θ = (6) ͛ ʚ0ʛ ƚ 0 ͕͘͢ ͛ ʚ0ʛ ƚ 0 : allows ͮ λ0 θͮ ̈́` Ɣ ʚ ͮ ͯʛ Ƭ ư Ƭ ư ͮͮͮ ƍͯͯͯ Ǝ isotropic lim |ǧ0| Ŵͤdiffusion lim for| ǧ0low|Ŵͤ gradient. ͙͗ͨ Ƙ 0 0 ͯ ͯ ͗ Ɣ ͗ Ɣ The eigenvectors of give the preferred local ͗ orientations, and the corresponding̈́` eigenvalues denote allowsƎ |ǧ0lim| ŴϦanisotropic͗ Ɣ |ǧ0lim |ŴϦdiffusion͗ Ɣ 0to ͕͘͢ preserve|ǧ0lim|ŴϦ discontinu͗ Ɣ 0: ities the local contrast along these directions. for the high gradient. The eigenvalues of are given by : ̈́` (7) In the coherence enhancing diffusion (CED) proposed by ͥ ͦ ͦ Weiker [4], the eigenvalues are assembled via: ͮ Ɣ ͦ ʚjͥͥ ƍ jͦͦ ƍ ǭʚjͥͥ Ǝ jͦͦ ʛ ƍ 4jͥͦ ʛ (8) ͥ ͦ ͦ (12) And the eigenvectorsͯ ͥͥ ( ͦͦ satisfyͥͥ :ͦͦ ͥͦ Ɣ ͦ ʚj ƍ j Ǝ ǭʚj Ǝ j ʛ ƍ 4j ʛ ͥ Ɣ ͗ͥ = ͮ, ͯʛ ͗ͥ ͚͝ ͮ Ɣ ͯ ͦ v ƥ ͥ ͥ v Ʃ ͦ%uv ͗ ƍ ʚ1 Ǝ ͗ ʛ exp ʠƎ ʚz~ͯzʛ ʡ ͙͙ͧ͠ v v v v ǯʚ%vv ͯ%uu ͮǭʚ%uu ͯ%vv ʛ ͮͨ%uv ʛ ͮͨ%uv (9) Where and .
Recommended publications
  • Scale Space Multiresolution Analysis of Random Signals
    Scale Space Multiresolution Analysis of Random Signals Lasse Holmstr¨om∗, Leena Pasanen Department of Mathematical Sciences, University of Oulu, Finland Reinhard Furrer Institute of Mathematics, University of Zurich,¨ Switzerland Stephan R. Sain National Center for Atmospheric Research, Boulder, Colorado, USA Abstract A method to capture the scale-dependent features in a random signal is pro- posed with the main focus on images and spatial fields defined on a regular grid. A technique based on scale space smoothing is used. However, where the usual scale space analysis approach is to suppress detail by increasing smoothing progressively, the proposed method instead considers differences of smooths at neighboring scales. A random signal can then be represented as a sum of such differences, a kind of a multiresolution analysis, each difference representing de- tails relevant at a particular scale or resolution. Bayesian analysis is used to infer which details are credible and which are just artifacts of random variation. The applicability of the method is demonstrated using noisy digital images as well as global temperature change fields produced by numerical climate prediction models. Keywords: Scale space smoothing, Bayesian methods, Image analysis, Climate research 1. Introduction In signal processing, smoothing is often used to suppress noise. An optimal level of smoothing is then of interest. In scale space analysis of noisy curves and images, instead of a single, in some sense “optimal” level of smoothing, a whole family of smooths is considered and each smooth is thought to provide information about the underlying object at a particular scale or resolution. The ∗Corresponding author ∗∗P.O.Box 3000, 90014 University of Oulu, Finland Tel.
    [Show full text]
  • Scale-Space Theory for Multiscale Geometric Image Analysis
    Tutorial Scale-Space Theory for Multiscale Geometric Image Analysis Bart M. ter Haar Romeny, PhD Utrecht University, the Netherlands [email protected] Introduction Multiscale image analysis has gained firm ground in computer vision, image processing and models of biological vision. The approaches however have been characterised by a wide variety of techniques, many of them chosen ad hoc. Scale-space theory, as a relatively new field, has been established as a well founded, general and promising multiresolution technique for image structure analysis, both for 2D, 3D and time series. The rather mathematical nature of many of the classical papers in this field has prevented wide acceptance so far. This tutorial will try to bridge that gap by giving a comprehensible and intuitive introduction to this field. We also try, as a mutual inspiration, to relate the computer vision modeling to biological vision modeling. The mathematical rigor is much relaxed for the purpose of giving the broad picture. In appendix A a number of references are given as a good starting point for further reading. The multiscale nature of things In mathematics objects have no scale. We are familiar with the notion of points, that really shrink to zero, lines with zero width. In mathematics are no metrical units involved, as in physics. Neighborhoods, like necessary in the definition of differential operators, are defined as taken into the limit to zero, so we can really speak of local operators. In physics objects live on a range of scales. We need an instrument to do an observation (our eye, a camera) and it is the range that this instrument can see that we call the scale range.
    [Show full text]
  • An Overview of Wavelet Transform Concepts and Applications
    An overview of wavelet transform concepts and applications Christopher Liner, University of Houston February 26, 2010 Abstract The continuous wavelet transform utilizing a complex Morlet analyzing wavelet has a close connection to the Fourier transform and is a powerful analysis tool for decomposing broadband wavefield data. A wide range of seismic wavelet applications have been reported over the last three decades, and the free Seismic Unix processing system now contains a code (succwt) based on the work reported here. Introduction The continuous wavelet transform (CWT) is one method of investigating the time-frequency details of data whose spectral content varies with time (non-stationary time series). Moti- vation for the CWT can be found in Goupillaud et al. [12], along with a discussion of its relationship to the Fourier and Gabor transforms. As a brief overview, we note that French geophysicist J. Morlet worked with non- stationary time series in the late 1970's to find an alternative to the short-time Fourier transform (STFT). The STFT was known to have poor localization in both time and fre- quency, although it was a first step beyond the standard Fourier transform in the analysis of such data. Morlet's original wavelet transform idea was developed in collaboration with the- oretical physicist A. Grossmann, whose contributions included an exact inversion formula. A series of fundamental papers flowed from this collaboration [16, 12, 13], and connections were soon recognized between Morlet's wavelet transform and earlier methods, including harmonic analysis, scale-space representations, and conjugated quadrature filters. For fur- ther details, the interested reader is referred to Daubechies' [7] account of the early history of the wavelet transform.
    [Show full text]
  • Scale-Equalizing Pyramid Convolution for Object Detection
    Scale-Equalizing Pyramid Convolution for Object Detection Xinjiang Wang,* Shilong Zhang∗, Zhuoran Yu, Litong Feng, Wayne Zhang SenseTime Research {wangxinjiang, zhangshilong, yuzhuoran, fenglitong, wayne.zhang}@sensetime.com Abstract 42 41 FreeAnchor C-Faster Feature pyramid has been an efficient method to extract FSAF D-Faster 40 features at different scales. Development over this method Reppoints mainly focuses on aggregating contextual information at 39 different levels while seldom touching the inter-level corre- L-Faster AP lation in the feature pyramid. Early computer vision meth- 38 ods extracted scale-invariant features by locating the fea- FCOS RetinaNet ture extrema in both spatial and scale dimension. Inspired 37 by this, a convolution across the pyramid level is proposed Faster Baseline 36 SEPC-lite in this study, which is termed pyramid convolution and is Two-stage detectors a modified 3-D convolution. Stacked pyramid convolutions 35 directly extract 3-D (scale and spatial) features and out- 60 65 70 75 80 85 90 performs other meticulously designed feature fusion mod- Time (ms) ules. Based on the viewpoint of 3-D convolution, an inte- grated batch normalization that collects statistics from the whole feature pyramid is naturally inserted after the pyra- Figure 1: Performance on COCO-minival dataset of pyramid mid convolution. Furthermore, we also show that the naive convolution in various single-stage detectors including RetinaNet [20], FCOS [38], FSAF [48], Reppoints [44], FreeAnchor [46]. pyramid convolution, together with the design of RetinaNet Reference points of two-stage detectors such as Faster R-CNN head, actually best applies for extracting features from a (Faster) [31], Libra Faster R-CNN (L-Faster) [29], Cascade Faster Gaussian pyramid, whose properties can hardly be satis- R-CNN (C-Faster) [1] and Deformable Faster R-CNN (D-Faster) fied by a feature pyramid.
    [Show full text]
  • Scale Normalized Image Pyramids with Autofocus for Object Detection
    1 Scale Normalized Image Pyramids with AutoFocus for Object Detection Bharat Singh, Mahyar Najibi, Abhishek Sharma and Larry S. Davis Abstract—We present an efficient foveal framework to perform object detection. A scale normalized image pyramid (SNIP) is generated that, like human vision, only attends to objects within a fixed size range at different scales. Such a restriction of objects’ size during training affords better learning of object-sensitive filters, and therefore, results in better accuracy. However, the use of an image pyramid increases the computational cost. Hence, we propose an efficient spatial sub-sampling scheme which only operates on fixed-size sub-regions likely to contain objects (as object locations are known during training). The resulting approach, referred to as Scale Normalized Image Pyramid with Efficient Resampling or SNIPER, yields up to 3× speed-up during training. Unfortunately, as object locations are unknown during inference, the entire image pyramid still needs processing. To this end, we adopt a coarse-to-fine approach, and predict the locations and extent of object-like regions which will be processed in successive scales of the image pyramid. Intuitively, it’s akin to our active human-vision that first skims over the field-of-view to spot interesting regions for further processing and only recognizes objects at the right resolution. The resulting algorithm is referred to as AutoFocus and results in a 2.5-5× speed-up during inference when used with SNIP. Code: https://github.com/mahyarnajibi/SNIPER Index Terms—Object Detection, Image Pyramids , Foveal vision, Scale-Space Theory, Deep-Learning. F 1 INTRODUCTION BJECT-detection is one of the most popular and widely O researched problems in the computer vision community, owing to its application to a myriad of industrial systems, such as autonomous driving, robotics, surveillance, activity-detection, scene-understanding and/or large-scale multimedia analysis.
    [Show full text]
  • Multi-Scale Edge Detection and Image Segmentation
    MULTI-SCALE EDGE DETECTION AND IMAGE SEGMENTATION Baris Sumengen, B. S. Manjunath ECE Department, UC, Santa Barbara 93106, Santa Barbara, CA, USA email: {sumengen,manj}@ece.ucsb.edu web: vision.ece.ucsb.edu ABSTRACT In this paper, we propose a novel multi-scale edge detection and vector field design scheme. We show that using multi- scale techniques edge detection and segmentation quality on natural images can be improved significantly. Our ap- proach eliminates the need for explicit scale selection and edge tracking. Our method favors edges that exist at a wide range of scales and localize these edges at finer scales. This work is then extended to multi-scale image segmentation us- (a) (b) (c) (d) ing our anisotropic diffusion scheme. Figure 1: Localized and clean edges using multiple scales. a) 1. INTRODUCTION Original image. b) Edge strengths at spatial scale σ = 1, c) Most edge detection algorithms specify a spatial scale at using scales from σ = 1 to σ = 4. d) at σ = 4. Edges are not which the edges are detected. Typically, edge detectors uti- well localized at σ = 4. lize local operators and the effective area of these local oper- ators define this spatial scale. The spatial scale usually corre- sponds to the level of smoothing of the image, for example, scales and generating a synthesis of these edges. On the the variance of the Gaussian smoothing. At small scales cor- other hand, it is desirable that the multi-scale information responding to finer image details, edge detectors find inten- is integrated to the edge detection at an earlier stage and the sity jumps in small neighborhoods.
    [Show full text]
  • Computer Vision: Edge Detection
    Edge Detection Edge detection Convert a 2D image into a set of curves • Extracts salient features of the scene • More compact than pixels Origin of Edges surface normal discontinuity depth discontinuity surface color discontinuity illumination discontinuity Edges are caused by a variety of factors Edge detection How can you tell that a pixel is on an edge? Profiles of image intensity edges Edge detection 1. Detection of short linear edge segments (edgels) 2. Aggregation of edgels into extended edges (maybe parametric description) Edgel detection • Difference operators • Parametric-model matchers Edge is Where Change Occurs Change is measured by derivative in 1D Biggest change, derivative has maximum magnitude Or 2nd derivative is zero. Image gradient The gradient of an image: The gradient points in the direction of most rapid change in intensity The gradient direction is given by: • how does this relate to the direction of the edge? The edge strength is given by the gradient magnitude The discrete gradient How can we differentiate a digital image f[x,y]? • Option 1: reconstruct a continuous image, then take gradient • Option 2: take discrete derivative (finite difference) How would you implement this as a cross-correlation? The Sobel operator Better approximations of the derivatives exist • The Sobel operators below are very commonly used -1 0 1 1 2 1 -2 0 2 0 0 0 -1 0 1 -1 -2 -1 • The standard defn. of the Sobel operator omits the 1/8 term – doesn’t make a difference for edge detection – the 1/8 term is needed to get the right gradient
    [Show full text]
  • Scale-Space and Edge Detection Using Anisotropic Diffusion
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 12. NO. 7. JULY 1990 629 Scale-Space and Edge Detection Using Anisotropic Diffusion PIETRO PERONA AND JITENDRA MALIK Abstracf-The scale-space technique introduced by Witkin involves generating coarser resolution images by convolving the original image with a Gaussian kernel. This approach has a major drawback: it is difficult to obtain accurately the locations of the “semantically mean- ingful” edges at coarse scales. In this paper we suggest a new definition of scale-space, and introduce a class of algorithms that realize it using a diffusion process. The diffusion coefficient is chosen to vary spatially in such a way as to encourage intraregion smoothing in preference to interregion smoothing. It is shown that the “no new maxima should be generated at coarse scales” property of conventional scale space is pre- served. As the region boundaries in our approach remain sharp, we obtain a high quality edge detector which successfully exploits global information. Experimental results are shown on a number of images. The algorithm involves elementary, local operations replicated over the Fig. 1. A family of l-D signals I(x, t)obtained by convolving the original image making parallel hardware implementations feasible. one (bottom) with Gaussian kernels whose variance increases from bot- tom to top (adapted from Witkin [21]). Zndex Terms-Adaptive filtering, analog VLSI, edge detection, edge enhancement, nonlinear diffusion, nonlinear filtering, parallel algo- rithm, scale-space. with the initial condition I(x, y, 0) = Zo(x, y), the orig- inal image. 1. INTRODUCTION Koenderink motivates the diffusion equation formula- HE importance of multiscale descriptions of images tion by stating two criteria.
    [Show full text]
  • Scalespaceviz: α-Scale Spaces in Practice
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Springer - Publisher Connector IMAGE PROCESSING, ANALYSIS, RECOGNITION, AND UNDERSTANDING ScaleSpaceViz: a-Scale Spaces in Practice¶ F. Kanters, L. Florack, R. Duits, B. Platel, and B. ter Haar Romeny Eindhoven University of Technology Department of Biomedical Engineering Den Dolech 2, Postbus 513 5600 MB Eindhoven, Netherlands e-mail: {F.M.W.Kanters, L.M.J.Florack, R.Duits, B.Platel, B.M.terHaarRomeny}@tue.nl Abstract—Kernels of the so-called α-scale space have the undesirable property of having no closed-form rep- resentation in the spatial domain, despite their simple closed-form expression in the Fourier domain. This obstructs spatial convolution or recursive implementation. For this reason an approximation of the 2D α-kernel in the spatial domain is presented using the well-known Gaussian kernel and the Poisson kernel. Experiments show good results, with maximum relative errors of less than 2.4%. The approximation has been successfully implemented in a program for visualizing α-scale spaces. Some examples of practical applications with scale space feature points using the proposed approximation are given. DOI: 10.1134/S1054661807010129 1. INTRODUCTION Witkin [2] and Koenderink [3] in the 1980s. A more recent introduction to scale space can be found in Scale is an essential parameter in computer vision, monograph by ter Haar Romeny [4]. It is, however, since it is an immediate consequence of the process of shown by Duits [5] that, for reasonable axioms, there observation, of measurements. Observations are always exists a complete class of kernels leading to the so- done by integrating some physical property with a called α-scale spaces, of which the Gaussian scale measurement device—for example, integration (over a space is a special case (α = 1).
    [Show full text]
  • Scale-Space and Edge Detection Using Anisotropic Diffusion
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 12. NO. 7. JULY 1990 629 Scale-Space and Edge Detection Using Anisotropic Diffusion PIETRO PERONA AND JITENDRA MALIK Abstracf-The scale-space technique introduced by Witkin involves generating coarser resolution images by convolving the original image with a Gaussian kernel. This approach has a major drawback: it is difficult to obtain accurately the locations of the “semantically mean- ingful” edges at coarse scales. In this paper we suggest a new definition of scale-space, and introduce a class of algorithms that realize it using a diffusion process. The diffusion coefficient is chosen to vary spatially in such a way as to encourage intraregion smoothing in preference to interregion smoothing. It is shown that the “no new maxima should be generated at coarse scales” property of conventional scale space is pre- served. As the region boundaries in our approach remain sharp, we obtain a high quality edge detector which successfully exploits global information. Experimental results are shown on a number of images. The algorithm involves elementary, local operations replicated over the Fig. 1. A family of l-D signals I(x, t)obtained by convolving the original image making parallel hardware implementations feasible. one (bottom) with Gaussian kernels whose variance increases from bot- tom to top (adapted from Witkin [21]). Zndex Terms-Adaptive filtering, analog VLSI, edge detection, edge enhancement, nonlinear diffusion, nonlinear filtering, parallel algo- rithm, scale-space. with the initial condition I(x, y, 0) = Zo(x, y), the orig- inal image. 1. INTRODUCTION Koenderink motivates the diffusion equation formula- HE importance of multiscale descriptions of images tion by stating two criteria.
    [Show full text]
  • Local Descriptors
    Local descriptors SIFT (Scale Invariant Feature Transform) descriptor • SIFT keypoints at locaon xy and scale σ have been obtained according to a procedure that guarantees illuminaon and scale invance. By assigning a consistent orientaon, the SIFT keypoint descriptor can be also orientaon invariant. • SIFT descriptor is therefore obtained from the following steps: – Determine locaon and scale by maximizing DoG in scale and in space (i.e. find the blurred image of closest scale) – Sample the points around the keypoint and use the keypoint local orientaon as the dominant gradient direc;on. Use this scale and orientaon to make all further computaons invariant to scale and rotaon (i.e. rotate the gradients and coordinates by the dominant orientaon) – Separate the region around the keypoint into subregions and compute a 8-bin gradient orientaon histogram of each subregion weigthing samples with a σ = 1.5 Gaussian SIFT aggregaon window • In order to derive a descriptor for the keypoint region we could sample intensi;es around the keypoint, but they are sensi;ve to ligh;ng changes and to slight errors in x, y, θ. In order to make keypoint es;mate more reliable, it is usually preferable to use a larger aggregaon window (i.e. the Gaussian kernel size) than the detec;on window. The SIFT descriptor is hence obtained from thresholded image gradients sampled over a 16x16 array of locaons in the neighbourhood of the detected keypoint, using the level of the Gaussian pyramid at which the keypoint was detected. For each 4x4 region samples they are accumulated into a gradient orientaon histogram with 8 sampled orientaons.
    [Show full text]
  • Anatomy of the SIFT Method Ives Rey Otero
    Anatomy of the SIFT method Ives Rey Otero To cite this version: Ives Rey Otero. Anatomy of the SIFT method. General Mathematics [math.GM]. École normale supérieure de Cachan - ENS Cachan, 2015. English. NNT : 2015DENS0044. tel-01226489 HAL Id: tel-01226489 https://tel.archives-ouvertes.fr/tel-01226489 Submitted on 9 Nov 2015 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Ecole´ Normale Sup´erieure de Cachan, France Anatomy of the SIFT method A dissertation presented by Ives Rey-Otero in fulfillment of the requirements for the degree of Doctor of Philosophy in the subject of Applied Mathematics Committee in charge Referees Coloma Ballester - Universitat Pompeu Fabra, ES Pablo Muse´ - Universidad de la Rep´ublica, UY Joachim Weickert - Universit¨at des Saarlandes, DE Advisors Jean-Michel Morel - ENS de Cachan, FR Mauricio Delbracio - Duke University, USA Examiners Patrick Perez´ - Technicolor Research, FR Fr´ed´eric Sur - Loria, FR ENSC-2015 603 September 2015 Version of November 5, 2015 at 12:27. Abstract of the Dissertation Ives Rey-Otero Anatomy of the SIFT method Under the direction of: Jean-Michel Morel and Mauricio Delbracio This dissertation contributes to an in-depth analysis of the SIFT method.
    [Show full text]