A Review Paper on Hough Transform and It's Applications in Image

Total Page:16

File Type:pdf, Size:1020Kb

A Review Paper on Hough Transform and It's Applications in Image ISSN(Online): 2319-8753 ISSN (Print) : 2347-6710 International Journal of Innovative Research in Science, Engineering and Technology (An ISO 3297: 2007 Certified Organization) Vol. 5, Special Issue 13, October 2016 A Review Paper on Hough Transform and it’s Applications in Image Processing Biswajit Sit 1, Md. Iqbal Quraishi 2 P.G. Student, Department of Computer Science Engineering, Kalyani Govt. Engg. College, Kalyani, Nadia, India1 Assistant Professor, Department of Information Technology, Kalyani Govt. Engg. College, Kalyani, Nadia, India2 ABSTRACT: The Hough transform is a feature extraction method that can be used in image analysis and digital image processing. The aim of this technique is to produce a computer vision system that can detect arbitrary shapes within a sample image. The main purpose of this method is finding imperfect instances of objects within a certain class of shapes by a voting procedure. The classical Hough transform was mainly introduced for the identification of lines in images, but later the Hough transform has been modified and extended to identify the positions of arbitrary shapes within an image, most commonly the extended version indulged itself in finding circles or ellipses. In that case appropriate parametric representation is needed. Nowadays there are a wide range of areas where the Hough Transform can be implemented successfully such as in medical visualization or in order to achieve high accuracy in face recognition etc. The characteristics of Pupil and Iris under Uncontrolled illumination can also be obtained by Hough Circle Transform. In Objective Spinal Motion Imaging Assessment system (OSMIA), it is required to locate marker that can be used in determining the positions of the vertebral bodies. The measurement of vertebral motion has been a challenge to the field of biomechanics for many years but now several automatic approaches to these problems have been developed. The Hough Transform has also been introduced in morphological image processing to detect and estimate the number of red blood cells in the blood sample image. KEYWORDS: Hough Transform, Line detection, Medical visualization, Pupil detection, Vertebrae detection, Red blood cell estimation. I. INTRODUCTION The Hough transform is a kind of technique used in image processing for extracting the features of an image. It finds the imperfect instances of an object within a group of shapes by a voting procedure. This procedure is done in a parameter space, where object candidates are obtained as local maxima in an accumulator space that is explicitly constructed by the algorithm for computing the Hough transform. The Hough transform was invented by Richard Duda and Peter Hart in the year 1972, and they named it a "generalized Hough transform"[1] after the related 1962 patent of Paul Hough [2]. The transform got popularity in 1981 by Dana H. Ballard through a journal article titled "Generalizing the Hough transform to detect arbitrary shapes". The most common case of using the Hough transform is the linear transform for detecting straight lines in an image. In the image space, the straight line can be described as y = mx + c where the parameter m represents slope of the line, and c represents the y-intercept. This form of representation is called the slope-intercept model of a straight line. The main idea behind using Hough Transform is considering the features of the straight line not as discrete image points (x1, y1), (x2, y2), etc., but in terms of slope-intercept model. In general, the straight line y = mx + c can be denoted as a point (b, c) in the parameter space where b represents the slope of the line and c represents the intercept. However, problems arises in case of vertical lines as they are represented as x = a and would give rise to unbounded values of the slope parameter m. Thus for this computational phenomenon, Duda and Hart proposed using of a different pair of parameters, denoted by r and θ (theta), for the lines in the Hough transform. These two values taken together as a whole define a point in the polar coordinate. Copyright to IJIRSET www.ijirset.com 206 ISSN(Online): 2319-8753 ISSN (Print) : 2347-6710 International Journal of Innovative Research in Science, Engineering and Technology (An ISO 3297: 2007 Certified Organization) Vol. 5, Special Issue 13, October 2016 Figure 1: Slope-intercept model of straight line The parameter r represents the algebraic distance between the line and the origin, while θ is the angle of the vector. If the line is located above the origin, θ is simply the angle of the vector from the origin to this closest point. Using this form, the equation of the line can be written as Which can be rearranged to r = xcos θ + ysin θ .It is therefore possible to associate with each line of the image a pair (r, θ) which is unique if θ ϵ [0, π) and r ϵ R, or if θ ϵ [0, 2π) and r≥0. The (r, θ) plane is sometimes referred to as Hough space for the set of straight lines in two dimensions. This representation makes this transform very similar to the Radon transform. For an arbitrary point on the image space, e.g. (x0, y0), the lines that go through this point are the pairs (r, θ) with r (θ) = x0cos θ + y0sin θ, Where r (the distance between the line and the origin) is determined by θ ϵ [0, π) .If r is required to be positive, then θ must vary in [0,2π). In other words, θ is the angle of the vector from the origin and this closest point (if r ≠ 0), or the angle of the vector orthogonal to the line and pointing to the half upper plane (if r=0). The lines that goes through (x0, y0) are the r (θ) = |x0cos θ + y0sin θ|. These representations leads to a sinusoidal curve in the (r, θ) plane [3], which is unique to that point. If the curves corresponding to the points are superimposed, then the location in the Hough space where they crosses each other corresponds to a line in the original image space that passes through both points. To be more general, a set of points that form a straight line will produce sinusoid curves that cross at the parameters for that line. Thus, the problem to detect collinear points can be converted to the problem of finding concurrent curves. II. DIFFERENT HOUGH TRANSFORMS A.LINE HOUGH TRANSFORM: A suitable equation for describing a set of lines in a parametric form is: x cosθ + ysinθ = r Where r is the length of the normal from the origin to this line and θ is the orientation of r with respect to the X-axis. A suitable way of describing this presentation is described in the figure 2 which shows an object taken as an example and the figure 3 shows the parametric representation of a line. Copyright to IJIRSET www.ijirset.com 207 ISSN(Online): 2319-8753 ISSN (Print) : 2347-6710 International Journal of Innovative Research in Science, Engineering and Technology (An ISO 3297: 2007 Certified Organization) Vol. 5, Special Issue 13, October 2016 Figure 2: gradients of the example object Figure 3: parametric representation of a line When an image is analyzed the (x, y) parameters represents the well known pixels that are selected for analyzing. The pair of θ and r used for parametric representation is inserted into an accumulator. Another approach to think about this fact is that all the lines that go through the point (x, y) are transformed into parametric space (θ, r) and then the relevant cell is increased by 1.Each and every single point present in the accumulator corresponds to certain line in the image. As this accumulator is discrete in nature it only consists of a set of all possible lines in R2. It also corresponds to a set of sinusoid curves which intersects in some points. The white portions in the figure 4 show the intersection area. Figure 4: Hough transformed image A very useful thing of this algorithm is the robustness of this approach against noise and gaps present in the input image. This technique can be useful in real life applications such as in analyzing ultrasound images which contains some obvious noise that may causes problems in analyzing the features. B. CIRCLE HOUGH TRANSFORM: To use the Hough Transform for detecting circles a suitable parametric representation is required. A circle can be represented as: (x-ax) 2 + (y-ay) 2 = R2 Where (ax, ay) denotes the center of the circle and R is the radius of the circle. A 3D accumulator array is also Copyright to IJIRSET www.ijirset.com 208 ISSN(Online): 2319-8753 ISSN (Print) : 2347-6710 International Journal of Innovative Research in Science, Engineering and Technology (An ISO 3297: 2007 Certified Organization) Vol. 5, Special Issue 13, October 2016 required for these three parameters used in representing the parametric equation. On the other hand if the radius of the circle is known in advance then the accumulator becomes two dimensional. There is a very close similarity between the circle Hough Transform [4] and line Hough Transform but the difference come into existence where every set pixel proposes every circle that satisfies (ax, ay, R). If there are many such pixels Proposes a certain circle or lie on the same circle a peak arises and it can be detected in the accumulator using thresholding. This circle detection method can also be modified to identify other irregular shaped objects (e.g. ellipse) that can be represented in a parametric form. In that case the algorithm turns into a problematic one as the complexity of finding the object get increased.
Recommended publications
  • Hough Transform, Descriptors Tammy Riklin Raviv Electrical and Computer Engineering Ben-Gurion University of the Negev Hough Transform
    DIGITAL IMAGE PROCESSING Lecture 7 Hough transform, descriptors Tammy Riklin Raviv Electrical and Computer Engineering Ben-Gurion University of the Negev Hough transform y m x b y m 3 5 3 3 2 2 3 7 11 10 4 3 2 3 1 4 5 2 2 1 0 1 3 3 x b Slide from S. Savarese Hough transform Issues: • Parameter space [m,b] is unbounded. • Vertical lines have infinite gradient. Use a polar representation for the parameter space Hough space r y r q x q x cosq + ysinq = r Slide from S. Savarese Hough Transform Each point votes for a complete family of potential lines: Each pencil of lines sweeps out a sinusoid in Their intersection provides the desired line equation. Hough transform - experiments r q Image features ρ,ϴ model parameter histogram Slide from S. Savarese Hough transform - experiments Noisy data Image features ρ,ϴ model parameter histogram Need to adjust grid size or smooth Slide from S. Savarese Hough transform - experiments Image features ρ,ϴ model parameter histogram Issue: spurious peaks due to uniform noise Slide from S. Savarese Hough Transform Algorithm 1. Image à Canny 2. Canny à Hough votes 3. Hough votes à Edges Find peaks and post-process Hough transform example http://ostatic.com/files/images/ss_hough.jpg Incorporating image gradients • Recall: when we detect an edge point, we also know its gradient direction • But this means that the line is uniquely determined! • Modified Hough transform: for each edge point (x,y) θ = gradient orientation at (x,y) ρ = x cos θ + y sin θ H(θ, ρ) = H(θ, ρ) + 1 end Finding lines using Hough transform
    [Show full text]
  • Scale Invariant Feature Transform (SIFT) Why Do We Care About Matching Features?
    Scale Invariant Feature Transform (SIFT) Why do we care about matching features? • Camera calibration • Stereo • Tracking/SFM • Image moiaicing • Object/activity Recognition • … Objection representation and recognition • Image content is transformed into local feature coordinates that are invariant to translation, rotation, scale, and other imaging parameters • Automatic Mosaicing • http://www.cs.ubc.ca/~mbrown/autostitch/autostitch.html We want invariance!!! • To illumination • To scale • To rotation • To affine • To perspective projection Types of invariance • Illumination Types of invariance • Illumination • Scale Types of invariance • Illumination • Scale • Rotation Types of invariance • Illumination • Scale • Rotation • Affine (view point change) Types of invariance • Illumination • Scale • Rotation • Affine • Full Perspective How to achieve illumination invariance • The easy way (normalized) • Difference based metrics (random tree, Haar, and sift, gradient) How to achieve scale invariance • Pyramids • Scale Space (DOG method) Pyramids – Divide width and height by 2 – Take average of 4 pixels for each pixel (or Gaussian blur with different ) – Repeat until image is tiny – Run filter over each size image and hope its robust How to achieve scale invariance • Scale Space: Difference of Gaussian (DOG) – Take DOG features from differences of these images‐producing the gradient image at different scales. – If the feature is repeatedly present in between Difference of Gaussians, it is Scale Invariant and should be kept. Differences Of Gaussians
    [Show full text]
  • Exploiting Information Theory for Filtering the Kadir Scale-Saliency Detector
    Introduction Method Experiments Conclusions Exploiting Information Theory for Filtering the Kadir Scale-Saliency Detector P. Suau and F. Escolano {pablo,sco}@dccia.ua.es Robot Vision Group University of Alicante, Spain June 7th, 2007 P. Suau and F. Escolano Bayesian filter for the Kadir scale-saliency detector 1 / 21 IBPRIA 2007 Introduction Method Experiments Conclusions Outline 1 Introduction 2 Method Entropy analysis through scale space Bayesian filtering Chernoff Information and threshold estimation Bayesian scale-saliency filtering algorithm Bayesian scale-saliency filtering algorithm 3 Experiments Visual Geometry Group database 4 Conclusions P. Suau and F. Escolano Bayesian filter for the Kadir scale-saliency detector 2 / 21 IBPRIA 2007 Introduction Method Experiments Conclusions Outline 1 Introduction 2 Method Entropy analysis through scale space Bayesian filtering Chernoff Information and threshold estimation Bayesian scale-saliency filtering algorithm Bayesian scale-saliency filtering algorithm 3 Experiments Visual Geometry Group database 4 Conclusions P. Suau and F. Escolano Bayesian filter for the Kadir scale-saliency detector 3 / 21 IBPRIA 2007 Introduction Method Experiments Conclusions Local feature detectors Feature extraction is a basic step in many computer vision tasks Kadir and Brady scale-saliency Salient features over a narrow range of scales Computational bottleneck (all pixels, all scales) Applied to robot global localization → we need real time feature extraction P. Suau and F. Escolano Bayesian filter for the Kadir scale-saliency detector 4 / 21 IBPRIA 2007 Introduction Method Experiments Conclusions Salient features X HD(s, x) = − Pd,s,x log2Pd,s,x d∈D Kadir and Brady algorithm (2001): most salient features between scales smin and smax P.
    [Show full text]
  • Hough Transform 1 Hough Transform
    Hough transform 1 Hough transform The Hough transform ( /ˈhʌf/) is a feature extraction technique used in image analysis, computer vision, and digital image processing.[1] The purpose of the technique is to find imperfect instances of objects within a certain class of shapes by a voting procedure. This voting procedure is carried out in a parameter space, from which object candidates are obtained as local maxima in a so-called accumulator space that is explicitly constructed by the algorithm for computing the Hough transform. The classical Hough transform was concerned with the identification of lines in the image, but later the Hough transform has been extended to identifying positions of arbitrary shapes, most commonly circles or ellipses. The Hough transform as it is universally used today was invented by Richard Duda and Peter Hart in 1972, who called it a "generalized Hough transform"[2] after the related 1962 patent of Paul Hough.[3] The transform was popularized in the computer vision community by Dana H. Ballard through a 1981 journal article titled "Generalizing the Hough transform to detect arbitrary shapes". Theory In automated analysis of digital images, a subproblem often arises of detecting simple shapes, such as straight lines, circles or ellipses. In many cases an edge detector can be used as a pre-processing stage to obtain image points or image pixels that are on the desired curve in the image space. Due to imperfections in either the image data or the edge detector, however, there may be missing points or pixels on the desired curves as well as spatial deviations between the ideal line/circle/ellipse and the noisy edge points as they are obtained from the edge detector.
    [Show full text]
  • Histogram of Directions by the Structure Tensor
    Histogram of Directions by the Structure Tensor Josef Bigun Stefan M. Karlsson Halmstad University Halmstad University IDE SE-30118 IDE SE-30118 Halmstad, Sweden Halmstad, Sweden [email protected] [email protected] ABSTRACT entity). Also, by using the approach of trying to reduce di- Many low-level features, as well as varying methods of ex- rectionality measures to the structure tensor, insights are to traction and interpretation rely on directionality analysis be gained. This is especially true for the study of the his- (for example the Hough transform, Gabor filters, SIFT de- togram of oriented gradient (HOGs) features (the descriptor scriptors and the structure tensor). The theory of the gra- of the SIFT algorithm[12]). We will present both how these dient based structure tensor (a.k.a. the second moment ma- are very similar to the structure tensor, but also detail how trix) is a very well suited theoretical platform in which to they differ, and in the process present a different algorithm analyze and explain the similarities and connections (indeed for computing them without binning. In this paper, we will often equivalence) of supposedly different methods and fea- limit ourselves to the study of 3 kinds of definitions of di- tures that deal with image directionality. Of special inter- rectionality, and their associated features: 1) the structure est to this study is the SIFT descriptors (histogram of ori- tensor, 2) HOGs , and 3) Gabor filters. The results of relat- ented gradients, HOGs). Our analysis of interrelationships ing the Gabor filters to the tensor have been studied earlier of prominent directionality analysis tools offers the possibil- [3], [9], and so for brevity, more attention will be given to ity of computation of HOGs without binning, in an algo- the HOGs.
    [Show full text]
  • The Hough Transform As a Tool for Image Analysis
    THE HOUGH TRANSFORM AS A TOOL FOR IMAGE ANALYSIS Josep Llad´os Computer Vision Center - Dept. Inform`atica. Universitat Aut`onoma de Barcelona. Computer Vision Master March 2003 Abstract The Hough transform is a widespread technique in image analysis. Its main idea is to transform the image to a parameter space where clusters or particular configurations identify instances of a shape under detection. In this chapter we overview some meaningful Hough-based techniques for shape detection, either parametrized or generalized shapes. We also analyze some approaches based on the straight line Hough transform able to detect particular structural properties in images. Some of the ideas of these approaches will be used in the following chapter to solve one of the goals of the present work. 1 Introduction The Hough transform was first introduced by Paul Hough in 1962 [4] with the aim of detecting alignments in T.V. lines. It became later the basis of a great number of image analysis applications. The Hough transform is mainly used to detect parametric shapes in images. It was first used to detect straight lines and later extended to other parametric models such as circumferences or ellipses, being finally generalized to any parametric shape [1]. The key idea of the Hough transform is that spatially extended patterns are transformed into a parameter space where they can be represented in a spatially compact way. Thus, a difficult global detection problem in the image space is reduced to an easier problem of peak detection in a parameter space. Ü; Ý µ A set of collinear image points ´ can be represented by the equation: ÑÜ Ò =¼ Ý (1) Ò where Ñ and are two parameters, the slope and intercept, which characterize the line.
    [Show full text]
  • Third Harmonic Generation Microscopy Zhang, Z
    VU Research Portal Third harmonic generation microscopy Zhang, Z. 2017 document version Publisher's PDF, also known as Version of record Link to publication in VU Research Portal citation for published version (APA) Zhang, Z. (2017). Third harmonic generation microscopy: Towards automatic diagnosis of brain tumors. General rights Copyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights. • Users may download and print one copy of any publication from the public portal for the purpose of private study or research. • You may not further distribute the material or use it for any profit-making activity or commercial gain • You may freely distribute the URL identifying the publication in the public portal ? Take down policy If you believe that this document breaches copyright please contact us providing details, and we will remove access to the work immediately and investigate your claim. E-mail address: [email protected] Download date: 05. Oct. 2021 Third harmonic generation microscopy: towards automatic diagnosis of brain tumors This thesis was reviewed by: prof.dr. J. Hulshof VU University Amsterdam prof.dr. J. Popp Jena University prof.dr. A.G.J.M. van Leeuwen Academic Medical Center prof.dr. M. van Herk The University of Manchester dr. I.H.M. van Stokkum VU University Amsterdam dr. P. de Witt Hamer VU University Medical Center © Copyright Zhiqing Zhang, 2017 ISBN: 978-94-6295-704-6 Printed in the Netherlands by Proefschriftmaken.
    [Show full text]
  • Visualize Yolo
    MSc Artificial Intelligence Master Thesis Open the black box: Visualize Yolo by Peter Heemskerk 11988797 August 18, 2020 36 EC credits autumn 2019 till summer 2020 Supervisor: Dr. Jan-Mark Geusebroek Assessor: Prof. Dr. Theo Gevers University of Amsterdam Contents 1 Introduction 2 1.1 Open the neural network black box . .2 1.2 Project Background . .2 1.2.1 Wet Cooling Towers . .2 1.2.2 The risk of Legionellosis . .2 1.2.3 The project . .3 2 Related work 4 2.1 Object Detection . .4 2.2 Circle detection using Hough transform . .4 2.3 Convolutional Neural Networks (ConvNet) . .5 2.4 You Only Look Once (Yolo) . .5 2.4.1 Yolo version 3 - bounding box and class prediction . .6 2.4.2 Yolo version 3 - object recognition at different scales . .7 2.4.3 Yolo version 3 - network architecture . .7 2.4.4 Feature Pyramid Networks . .7 2.5 The Black Box Explanation problem . .9 2.6 Network Visualization . 10 3 Approach 10 3.1 Aerial imagery dataset . 10 3.2 Yolo version 3 . 12 3.2.1 Pytorch Implementation . 12 3.2.2 Tuning approach . 12 3.3 Evaluation . 12 3.3.1 Training, test and validation sets . 12 3.3.2 Evaluation Metrics . 12 3.4 Network Visualization . 13 3.4.1 Introduction . 13 3.4.2 Grad-CAM . 13 3.4.3 Feature maps . 14 4 Experiment 15 4.1 Results . 15 4.1.1 Hough Transform prediction illustration . 15 4.1.2 Yolo Prediction illustration . 16 4.1.3 Yolo Tuning . 17 4.1.4 Yolo Validation .
    [Show full text]
  • Line Detection by Hough Transformation
    Line Detection by Hough transformation 09gr820 April 20, 2009 1 Introduction When images are to be used in different areas of image analysis such as object recognition, it is important to reduce the amount of data in the image while preserving the important, characteristic, structural information. Edge detection makes it possible to reduce the amount of data in an image considerably. However the output from an edge detector is still a image described by it’s pixels. If lines, ellipses and so forth could be defined by their characteristic equations, the amount of data would be reduced even more. The Hough transform was originally developed to recognize lines [5], and has later been generalized to cover arbitrary shapes [3] [1]. This worksheet explains how the Hough transform is able to detect (imperfect) straight lines. 2 The Hough Space 2.1 Representation of Lines in the Hough Space Lines can be represented uniquely by two parameters. Often the form in Equation 1 is used with parameters a and b. y = a · x + b (1) This form is, however, not able to represent vertical lines. Therefore, the Hough transform uses the form in Equation 2, which can be rewritten to Equation 3 to be similar to Equation 1. The parameters θ and r is the angle of the line and the distance from the line to the origin respectively. r = x · cos θ + y · sin θ ⇔ (2) cos θ r y = − · x + (3) sin θ sin θ All lines can be represented in this form when θ ∈ [0, 180[ and r ∈ R (or θ ∈ [0, 360[ and r ≥ 0).
    [Show full text]
  • Design of a Video Processing Algorithm for Detection of a Soccer Ball with Arbitrary Color Pattern
    Design of a video processing algorithm for detection of a soccer ball with arbitrary color pattern R. Woering DCT 2009.017 Traineeship report Coach(es): Ir. A.J. den Hamer Supervisor: prof.dr.ir. M. Steinbuch Technische Universiteit Eindhoven Department Mechanical Engineering Dynamics and Control Technology Group Eindhoven, March, 2009 Contents 1 Introduction 2 2 Literature 4 3 Basics of image processing 5 3.1 YUVandRGBcolorspaces ............................ ... 5 3.2 LinearFilters ................................... .... 7 Averagingfiltering .................................. .. 8 Gaussianlow-passfilter .. .. .. .. .. .. .. .. .. .. .. .. .. ... 8 LaplacianofGaussianfilter(LoG) . ..... 8 Unsharpfilter....................................... 9 3.3 Edgedetection ................................... ... 9 Cannymethod ...................................... 10 Sobelmethod....................................... 11 4 Circle Hough Transform (CHT) 12 4.1 Extraballcheck .................................. ... 12 5 Matlab 17 6 OpenCV 22 6.1 OpenCVreal-timetesting . ...... 23 7 Conclusion and Recommendations 25 References 28 1 1 Introduction This research is performed within the RoboCup project at the TU/e. RoboCup is an international joint project to promote A.I. (Artificial Intelligence), robotics and related fields. The idea is to perform research in the field of autonomous robots that play football by adapted FIFA rules. The goal is to play with humanoid robots against the world champion football of 2050 and hopefully win. Every year new challenges are set to force research and development to make it possible to play against humans in 2050. An autonomous mobile robot is a robot that is provided with the ability to take decisions on its own without interference of humans and work in a nondeterministic environment. A very important part of the development of autonomous robots is the real-time video processing, which is used to recognize the object in its surroundings.
    [Show full text]
  • Low-Level Vision Tutorial 1
    Low-level vision tutorial 1 complexity and sophistication – so much so that Feature detection and sensitivity the problems of low-level vision are often felt to Image filters and morphology be unimportant and are forgotten. Yet it remains Robustness of object location the case that information that is lost at low level Validity and accuracy in shape analysis Low-level vision – a tutorial is never regained, while distortions that are Scale and affine invariance. introduced at low level can cause undue trouble at higher levels [1]. Furthermore, image Professor Roy Davies In what follows, references are given for the acquisition is equally important. Thus, simple topics falling under each of these categories, and measures to arrange suitable lighting can help to for other topics that could not be covered in make the input images easier and less ambiguous Overview and further reading depth in the tutorial. to interpret, and can result in much greater reliability and accuracy in applications such as 3. Feature detection and sensitivity automated inspection [1]. Nevertheless, in applications such as surveillance, algorithms General [1]. Abstract should be made as robust as possible so that the Edge detection [2, 3, 4]. vagaries of ambient illumination are rendered Line segment detection [5, 6, 7]. This tutorial aims to help those with some relatively unimportant. Corner and interest point detection [1, 8, 9]. experience of vision to obtain a more in-depth General feature mask design [10, 11]. understanding of the problems of low-level 2. Low-level vision The value of thresholding [1, 12, 13]. vision. As it is not possible to cover everything in the space of 90 minutes, a carefully chosen This tutorial is concerned with low-level vision, 4.
    [Show full text]
  • Detection and Localization of Spherical Markers in Photographs
    Proceedings of the Croatian Computer Vision Workshop, Year 4 October 11, 2016, Osijek, Croatia Detection and Localization of Spherical Markers in Photographs Josip Tomurad Marko Subašić Faculty of Electrical Engineering and Computing Faculty of Electrical Engineering and Computing University of Zagreb University of Zagreb Zagreb, Croatia Zagreb, Croatia e-mail: [email protected] e-mail: [email protected] Abstract - This paper presents two solutions for detection II. HOUGH TRANSFORM and localization of spherical markers in photographs. Hough transform is a technique of feature extraction that The proposed solutions enable precise detection and is used in image analysis, computer vision, and digital localization in sub millimeter range. High precision image processing. The purpose of this technique is finding localization is required for brain surgery, and presented imperfect instances of objects from a certain class of shapes research effort is part of a project of developing and by application of voting. The technique was originally used deploying a robotic system for neurosurgical for line detection in images, and was later expanded to applications. Two algorithms for Hough transform using recognition of various kinds of shapes, most commonly several edge detection algorithms are proposed, and ellipses and circles. their results compared and analyzed. Results are obtained for both NIR and visible spectrum images, and Hough transform uses shape edges as its input so first required high precision is achieved in both domains. step is finding edge pixels in an image. Then each edge pixel votes in a Hough parameter space in a pattern that Keywords – project RONNA, image processing, circle describes potential shape of interest.
    [Show full text]