Contributions to a Fast and Robust Object Recognition in Images Jérôme Revaud

Total Page:16

File Type:pdf, Size:1020Kb

Contributions to a Fast and Robust Object Recognition in Images Jérôme Revaud Contributions to a fast and robust object recognition in images Jérôme Revaud To cite this version: Jérôme Revaud. Contributions to a fast and robust object recognition in images. Other [cs.OH]. INSA de Lyon, 2011. English. NNT : 2011ISAL0042. tel-00694442 HAL Id: tel-00694442 https://tel.archives-ouvertes.fr/tel-00694442 Submitted on 4 May 2012 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Numéro d’ordre : 2011ISAL0042 Année 2011 Institut National des Sciences Appliquées de Lyon Laboratoire d’InfoRmatique en Image et Systèmes d’information École Doctorale Informatique et Mathématiques de Lyon Thèse de l’Université de Lyon Présentée en vue d’obtenir le grade de Docteur, spécialité Informatique par Jérôme Revaud Contributions to a Fast and Robust Object Recognition in Images Soutenance prévue le 27 mai 2011 devant le jury composé de : M. Patrick Gros Directeur de recherche, INRIA Rennes Rapporteur M. Frédéric Jurie Professeur, Université de Caen Rapporteur M. Vincent Lepetit Senior Researcher, EPFL Examinateur M. Jean Ponce Professeur, INRIA Paris Examinateur M. Atilla Baskurt Professeur, INSA Lyon Directeur M. Yasuo Ariki Professeur, Université de Kobe Co-directeur M. Guillaume Lavoué Maître de Conférences, INSA Lyon Co-encadrant Laboratoire d’InfoRmatique en Image et Systèmes d’information UMR 5205 CNRS - INSA de Lyon - Bât. Jules Verne 69621 Villeurbanne cedex - France Tel: +33 (0)4 72 43 60 97 - Fax: +33 (0)4 72 43 71 17 Cette thèse est accessible à l'adresse : http://theses.insa-lyon.fr/publication/2011ISAL0042/these.pdf © [J. Revaud], [2011], INSA de Lyon, tous droits réservés Cette thèse est accessible à l'adresse : http://theses.insa-lyon.fr/publication/2011ISAL0042/these.pdf © [J. Revaud], [2011], INSA de Lyon, tous droits réservés Abstract Object recognition in images is a growing field. Since several years, the emergence of invariant interest points such as SIFT [Low01] has enabled rapid and effective systems for the recognition of instances of specific objects as well as classes of objects (e.g. using the bag-of-words model). However, our experiments on the recognition of specific object instances have shown that under realistic conditions of use (e.g. the presence of various noises such as blur, poor lighting, low resolution cameras, etc.) progress remains to be done in terms of recall: despite the low rate of false positives, too few actual instances are detected regardless of the system (RANSAC, votes / Hough ...). In this presentation, we first present a contribution to overcome this problem of robustness for the recogni- tion of object instances, then we straightly extend this contribution to the detection and localization of classes of objects. Initially, we have developed a method inspired by graph matching to address the problem of fast recognition of instances of specific objects in noisy conditions. This method allows to easily combine any types of local features (eg contours, textures ...) less affected by noise than keypoints, while bypassing the normalization problem and without penalizing too much the detection speed. In this approach, the detection system consists of a set of cascades of micro-classifiers trained beforehand. Each micro-classifier is responsible for comparing the test image locally and from a certain point of view (e.g. as contours, or textures ...) to the same area in the model image. The cascades of micro- classifiers can therefore recognize different parts of the model in a robust manner (only the most effective cascades are selected during learning). Finally, a probabilistic model that combines those partial detections infers global detections. Unlike other methods based on a global rigid transformation, our approach is robust to complex deformations such as those due to perspective or those non-rigid inherent to the model itself (e.g. a face, a flexible magazine). Our experiments on several datasets have showed the relevance of our approach. It is overall slightly less robust to occlusion than existing approaches, but it produces better performances in noisy conditions. In a second step, we have developed an approach for detecting classes of objects in the same spirit as the bag-of-visual-words model. For this we use our cascaded micro- classifiers to recognize visual words more distinctive than the classical words simply iii Cette thèse est accessible à l'adresse : http://theses.insa-lyon.fr/publication/2011ISAL0042/these.pdf © [J. Revaud], [2011], INSA de Lyon, tous droits réservés based on visual dictionaries (like Csurka et al. [CDF∗04] or Lazebnik et al. [LSP05]). Training is divided into two parts: First, we generate cascades of micro-classifiers for recognizing local parts of the model pictures and then in a second step, we use a clas- sifier to model the decision boundary between images of class and those of non-class. This classifier bases its decision on a vector counting the outputs of each binary micro- classifier. This vector is extremely sparse and a simple classifier such as Real-Adaboost manages to produce a system with good performances (this type of classifier is similar in fact to the subgraph membership kernel). In particular, we show that the association of classical visual words (from keypoints patches) and our disctinctive words results in a significant improvement. The computation time is generally quite low, given the struc- ture of the cascades that minimizes the detection time and the form of the classifier is extremely fast to evaluate. Keywords: Specific object recognition, class object recognition, graph matching, cas- cades, optimization, mobile robotic. iv Cette thèse est accessible à l'adresse : http://theses.insa-lyon.fr/publication/2011ISAL0042/these.pdf © [J. Revaud], [2011], INSA de Lyon, tous droits réservés Contents Abstract iii Contents v List of Figures ix List of Tables xiii List of Algorithms xv 1 Introduction 1 1.1 A Few Preliminary Words . 1 1.2 Application Field . 2 1.3 A Short Definition of Object Recognition Terms . 4 1.4 Outlines . 6 2 Survey on Object Recognition 9 2.1 A Glance at Object Recognition . 11 2.2 Low-level Features . 12 2.2.1 Dense features . 14 2.2.1.1 Convolution-based features . 14 2.2.1.2 Non-linear features . 16 2.2.2 Sparse features . 16 2.2.2.1 Edges . 17 2.2.2.2 Keypoints . 18 2.2.2.3 Regions . 20 2.2.3 Histogram-based features . 20 2.2.3.1 Local descriptors . 20 2.3 Specific Object Recognition . 23 2.3.1 Using global features . 23 2.3.2 Using local features . 24 2.3.2.1 Rigid matching . 25 v Cette thèse est accessible à l'adresse : http://theses.insa-lyon.fr/publication/2011ISAL0042/these.pdf © [J. Revaud], [2011], INSA de Lyon, tous droits réservés 2.3.2.2 Non-rigid matching . 29 2.4 Class Object Recognition . 33 2.4.1 Feature spaces for class object recognition . 34 2.4.2 Detection schemes . 36 3 Cascaded Multi-feature Incomplete Graph Matching For 3D Specific Object Recognition 41 3.1 Introduction and Motivations . 43 3.1.1 The feature combination problem . 44 3.1.2 Outlines of the proposed method . 46 3.1.3 Related works . 46 3.2 Useful notation . 49 3.3 Used Features . 49 3.3.1 Keypoints . 50 3.3.2 Edges . 51 3.3.3 Textures . 52 3.4 Algorithm Description . 52 3.4.1 The prototype graphs . 54 3.4.2 The detection lattice . 55 3.4.3 Aggregate position . 57 3.4.4 Aggregate recognition . 57 3.4.5 Clustering of detected aggregates . 60 3.4.6 Probabilistic model for clusters of hypothesis . 62 3.5 How to build the detection lattice . 66 3.5.1 Algorithm inputs . 67 3.5.2 Iterative pruning of the lattice . 67 3.5.3 Learning the micro-classifier thresholds . 69 3.5.4 Ranking of the aggregates . 69 3.5.5 Discretization of the training image into parts . 72 3.6 Conclusion . 73 4 Evaluation of Our Contribution For Specific Object Detection 75 4.1 Discussion about the evaluation . 77 4.1.1 Test datasets . 77 4.1.2 Evaluation metrics . 79 4.2 Preliminary training . 80 4.2.1 Learning the subclassifier thresholds . 80 4.2.2 Other kernel parameters . 83 4.3 The CS17 dataset . 83 4.3.1 Parameter Tuning . 84 4.3.2 Comparative experiments . 87 4.3.3 Discussion . 93 4.4 The ETHZ toys dataset . 95 vi Cette thèse est accessible à l'adresse : http://theses.insa-lyon.fr/publication/2011ISAL0042/these.pdf © [J. Revaud], [2011], INSA de Lyon, tous droits réservés 4.5 The Rothganger dataset . 98 4.6 Conclusion . 102 5 Extension of the Multi-feature Incomplete Graph Matching to Recognition of Class Objects 103 5.1 Introduction . 105 5.1.1 Method overview . 105 5.1.2 Related works . 107 5.1.3 Chapter outline . 109 5.2 Method Description . 110 5.2.1 Features used . 111 5.2.2 Window classification . 114 5.2.3 Optimization for training the classifier . 115 5.2.4 Optimization for detection speed . 117 5.3 Modifications to the original lattice . 117 5.3.1 Rotation variance . 118 5.3.2 Recognition procedure for the lattice . 118 5.3.3 Training procedure for the lattice . 120 5.4 Conclusion . 123 6 Evaluation of Our Contribution For Class Object Detection 125 6.1 Introduction .
Recommended publications
  • 3D Object Modeling and Recognition Using Local Affine-Invariant Image Descriptors and Multi-View Spatial Constraints
    3D Object Modeling and Recognition Using Local Affine-Invariant Image Descriptors and Multi-View Spatial Constraints Fred Rothganger ([email protected]) Svetlana Lazebnik ([email protected]) Department of Computer Science and Beckman Institute University of Illinois at Urbana-Champaign, Urbana, IL 61801, USA Cordelia Schmid ([email protected]) INRIA Rhone-Alpesˆ 665, Avenue de l’Europe, 38330 Montbonnot, France Jean Ponce ([email protected]) Department of Computer Science and Beckman Institute University of Illinois at Urbana-Champaign, Urbana, IL 61801, USA Abstract. This article introduces a novel representation for three-dimensional (3D) objects in terms of local affine-invariant descriptors of their images and the spatial relationships between the corresponding surface patches. Geometric constraints associated with different views of the same patches under affine projection are combined with a normalized representation of their appearance to guide matching and reconstruction, allowing the acquisition of true 3D affine and Euclidean models from multiple unregistered images, as well as their recognition in photographs taken from arbitrary viewpoints. The proposed approach does not require a separate segmentation stage, and it is applicable to highly cluttered scenes. Modeling and recognition results are presented. Keywords: Three-dimensional object recognition, image-based modeling, affine-invariant image descriptors, multi- view geometry. 1. Introduction This article addresses the problem of recognizing three-dimensional (3D) objects
    [Show full text]
  • View Matching with Blob Features
    View Matching with Blob Features Per-Erik Forssen´ and Anders Moe Computer Vision Laboratory Department of Electrical Engineering Linkoping¨ University, Sweden Abstract mographic transformation that is not part of the affine trans- formation. This means that our results are relevant for both This paper introduces a new region based feature for ob- wide baseline matching and 3D object recognition. ject recognition and image matching. In contrast to many A somewhat related approach to region based matching other region based features, this one makes use of colour is presented in [1], where tangents to regions are used to de- in the feature extraction stage. We perform experiments on fine linear constraints on the homography transformation, the repeatability rate of the features across scale and incli- which can then be found through linear programming. The nation angle changes, and show that avoiding to merge re- connection is that a line conic describes the set of all tan- gions connected by only a few pixels improves the repeata- gents to a region. By matching conics, we thus implicitly bility. Finally we introduce two voting schemes that allow us match the tangents of the ellipse-approximated regions. to find correspondences automatically, and compare them with respect to the number of valid correspondences they 2. Blob features give, and their inlier ratios. We will make use of blob features extracted using a clus- tering pyramid built using robust estimation in local image regions [2]. Each extracted blob is represented by its aver- 1. Introduction age colour pk,areaak, centroid mk, and inertia matrix Ik. I.e.
    [Show full text]
  • Object Detection and Pose Estimation from RGB and Depth Data for Real-Time, Adaptive Robotic Grasping
    Object Detection and Pose Estimation from RGB and Depth Data for Real-time, Adaptive Robotic Grasping Shuvo Kumar Paul1, Muhammed Tawfiq Chowdhury1, Mircea Nicolescu1, Monica Nicolescu1, David Feil-Seifer1 Abstract—In recent times, object detection and pose estimation to the changing environment while interacting with their have gained significant attention in the context of robotic vision surroundings, and a key component of this interaction is the applications. Both the identification of objects of interest as well reliable grasping of arbitrary objects. Consequently, a recent as the estimation of their pose remain important capabilities in order for robots to provide effective assistance for numerous trend in robotics research has focused on object detection and robotic applications ranging from household tasks to industrial pose estimation for the purpose of dynamic robotic grasping. manipulation. This problem is particularly challenging because However, identifying objects and recovering their poses are of the heterogeneity of objects having different and potentially particularly challenging tasks as objects in the real world complex shapes, and the difficulties arising due to background are extremely varied in shape and appearance. Moreover, clutter and partial occlusions between objects. As the main contribution of this work, we propose a system that performs cluttered scenes, occlusion between objects, and variance in real-time object detection and pose estimation, for the purpose lighting conditions make it even more difficult. Additionally, of dynamic robot grasping. The robot has been pre-trained to the system needs to be sufficiently fast to facilitate real-time perform a small set of canonical grasps from a few fixed poses robotic tasks.
    [Show full text]
  • 3D Object Modeling and Recognition from Photographs and Image Sequences
    3D Object Modeling and Recognition from Photographs and Image Sequences Fred Rothganger1, Svetlana Lazebnik1, Cordelia Schmid2, and Jean Ponce1 1 Department of Computer Science and Beckman Institute University of Illinois at Urbana-Champaign, Urbana, IL 61801, USA {rothgang,slazebni,jponce}@uiuc.edu 2 INRIA Rhˆone-Alpes 665, Avenue de l’Europe, 38330 Montbonnot, France [email protected] Abstract. This chapter proposes a representation of rigid three-dimensional (3D) objects in terms of local affine-invariant descriptors of their images and the spatial relationships between the corresponding surface patches. Geometric constraints associated with different views of the same patches under affine projection are combined with a normalized representation of their appearance to guide the matching process involved in object mod- eling and recognition tasks. The proposed approach is applied in two do- mains: (1) Photographs — models of rigid objects are constructed from small sets of images and recognized in highly cluttered shots taken from arbitrary viewpoints. (2) Video — dynamic scenes containing multiple moving objects are segmented into rigid components, and the resulting 3D models are directly matched to each other, giving a novel approach to video indexing and retrieval. 1 Introduction Traditional feature-based geometric approaches to three-dimensional (3D) object recognition — such as alignment [13, 19] or geometric hashing [15] — enumerate various subsets of geometric image features before using pose consistency con- straints to confirm or discard competing match hypotheses. They largely ignore the rich source of information contained in the image brightness and/or color pattern, and thus typically lack an effective mechanism for selecting promis- ing matches.
    [Show full text]
  • Computer Vision Based Human Detection Md Ashikur
    Computer Vision Based Human Detection Md Ashikur. Rahman To cite this version: Md Ashikur. Rahman. Computer Vision Based Human Detection. International Journal of Engineer- ing and Information Systems (IJEAIS), 2017, 1 (5), pp.62 - 85. hal-01571292 HAL Id: hal-01571292 https://hal.archives-ouvertes.fr/hal-01571292 Submitted on 2 Aug 2017 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. International Journal of Engineering and Information Systems (IJEAIS) ISSN: 2000-000X Vol. 1 Issue 5, July– 2017, Pages: 62-85 Computer Vision Based Human Detection Md. Ashikur Rahman Dept. of Computer Science and Engineering Shaikh Burhanuddin Post Graduate College Under National University, Dhaka, Bangladesh [email protected] Abstract: From still images human detection is challenging and important task for computer vision-based researchers. By detecting Human intelligence vehicles can control itself or can inform the driver using some alarming techniques. Human detection is one of the most important parts in image processing. A computer system is trained by various images and after making comparison with the input image and the database previously stored a machine can identify the human to be tested. This paper describes an approach to detect different shape of human using image processing.
    [Show full text]
  • Local Feature Filtering for 3D Object Recognition
    Local Feature Filtering for 3D Object Recognition Md. Kamrul Hasan∗, Liang Chen†, Christopher J. Pal∗ and Charles Brown† ∗ Ecole Polytechnique de Montreal, QC, Canada E-mail: [email protected], [email protected], Tel: +1(514) 340-5121,x.7110 † University of Northern British Columbia, BC, Canada E-mail: [email protected], [email protected] Tel: +1(250) 960-5838 Abstract—Three filtering/sampling approaches for local fea- by the Bag of Words (BOW) model from textual information tures, namely Random sampling, Concentrated sampling, and retrieval, computer vision researchers have induced it as Bag Inversely concentrated sampling are proposed and investigated. Of Visual Words(BOVW) model[13,14], where visual features While local features characterize an image with distinctive key- points, the filtering/sampling analogies have a two-fold goal: (i) are quantized, the code-words are generated, and named as compressing an object definition, and (ii) reducing the classifier visual words. These Visual Words have been successfully learning bias, that is induced by the variance of key-point tested for feature matching in 2D object images. numbers among object classes. The proposed methodologies have In this work, we have proposed three new local image fea- been tested for 3D object recognition on the Columbia Object ture filtering methodologies that are based on some very sim- Library (COIL-20) dataset, where the Support Vector Machines (SVMs) is applied as a classification tool and winner takes all ple assumptions. Following these filtering approaches, three principle is used for inferences. The proposed approaches have object recognition models have been formulated, which are achieved promising performance with 100% recognition rate for described in section two.
    [Show full text]
  • Object Recognition Using Hough-Transform Clustering of SURF Features
    Object Recognition Using Hough-transform Clustering of SURF Features Viktor Seib, Michael Kusenbach, Susanne Thierfelder, Dietrich Paulus Active Vision Group (AGAS), University of Koblenz-Landau Universit¨atsstr. 1, 56070 Koblenz, Germany {vseib, mkusenbach, thierfelder, paulus}@uni-koblenz.de http://homer.uni-koblenz.de Abstract—This paper proposes an object recognition approach important algorithm parameters. Finally, we provide an open intended for extracting, analyzing and clustering of features from source implementation of our approach as a Robot Operating RGB image views from given objects. Extracted features are System (ROS) package that can be used with any robot capable matched with features in learned object models and clustered in of interfacing ROS. The software is available at [3]. Hough-space to find a consistent object pose. Hypotheses for valid poses are verified by computing a homography from detected In Sec.II related approaches are briefly presented. These features. Using that homography features are back projected approaches either use similar techniques or are used by other onto the input image and the resulting area is checked for teams competing in the RoboCup@Home league. Hereafter, possible presence of other objects. This approach is applied by Sec.III presents our approach in great detail, an evaluation our team homer[at]UniKoblenz in the RoboCup[at]Home league. follows in Sec.IV. Finally, Sec.V concludes this paper. Besides the proposed framework, this work offers the computer vision community with online programs available as open source software. II. RELATED WORK Similar to our approach, Leibe and Schiele [4] use the I. INTRODUCTION Hough-space for creating hypotheses for the object pose.
    [Show full text]
  • 3D Object Processing and Image Processing by Numerical Methods Abdul Rahman El Sayed
    3D object processing and Image processing by numerical methods Abdul Rahman El Sayed To cite this version: Abdul Rahman El Sayed. 3D object processing and Image processing by numerical methods. General Mathematics [math.GM]. Normandie Université, 2018. English. NNT : 2018NORMLH19. tel- 01951684 HAL Id: tel-01951684 https://tel.archives-ouvertes.fr/tel-01951684 Submitted on 11 Dec 2018 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. THESE Pour obtenir le diplôme de doctorat Spécialité : Mathématiques Appliquées et Informatique Préparée au sein de « Université Le Havre Normandie - LMAH » Traitement des objets 3D et images par les méthodes numériques sur graphes Présentée et soutenue par Abdul Rahman EL SAYED Thèse soutenue publiquement le 24 octobre 2018 devant le jury composé de Professeur, Université du Littoral - M. Cyril FONLUPT Rapporteur Côte d'Opale (ULCO) Professeur, Université de M. Abderrafiaa KOUKAM Rapporteur Technologie de Belfort-Montbéliard Maître de Conférences, Université Mme Marianne DE BOYSSON Examinatrice Le Havre Normandie Professeur, Université du Littoral - M. Henri BASSON Examinateur Côte d'Opale (ULCO) Maître de Conférences, Beirut Arab M. Abdallah EL CHAKIK Examinateur University, Lebanon Maître de Conférences, Université M. Hassan ALABBOUD Co-encadreur de thèse Libanaise, Tripoli – Liban Professeur – Université Le Havre M.
    [Show full text]
  • 2D/3D Object Recognition and Categorization Approaches for Robotic Grasping Nabila Zrira, Mohamed Hannat, El Houssine Bouyakhf, Haris Ahmad Khan
    2D/3D Object Recognition and Categorization Approaches for Robotic Grasping Nabila Zrira, Mohamed Hannat, El Houssine Bouyakhf, Haris Ahmad Khan To cite this version: Nabila Zrira, Mohamed Hannat, El Houssine Bouyakhf, Haris Ahmad Khan. 2D/3D Object Recogni- tion and Categorization Approaches for Robotic Grasping. Advances in Soft Computing and Machine Learning in Image Processing, 730, pp.567-593, 2017, 978-3-319-63754-9. hal-01676387 HAL Id: hal-01676387 https://hal.archives-ouvertes.fr/hal-01676387 Submitted on 5 Jan 2018 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. 2D/3D Object Recognition and Categorization Approaches for Robotic Grasping Nabila Zrira, Mohamed Hannat, El Houssine Bouyakhf and Haris Ahmad Khan Abstract Object categorization and manipulation are critical tasks for a robot to operate in the household environment. In this paper, we propose new methods for visual recognition and categorization. We describe 2D object database and 3D point clouds with 2D/3D local descriptors which we quantify with the k-means clustering algorithm for obtaining the Bag of Words (BOW). Moreover, we develop a new global descriptor called VFH-Color that combines the original version of Viewpoint Feature Histogram (VFH) descriptor with the color quantization histogram, thus adding the appearance information that improves the recognition rate.
    [Show full text]
  • Distinctive Image Features from Scale-Invariant Keypoints
    Distinctive Image Features from Scale-Invariant Keypoints David G. Lowe Computer Science Department University of British Columbia Vancouver, B.C., Canada [email protected] January 5, 2004 Abstract This paper presents a method for extracting distinctive invariant features from images that can be used to perform reliable matching between different views of an object or scene. The features are invariant to image scale and rotation, and are shown to provide robust matching across a a substantial range of affine dis- tortion, change in 3D viewpoint, addition of noise, and change in illumination. The features are highly distinctive, in the sense that a single feature can be cor- rectly matched with high probability against a large database of features from many images. This paper also describes an approach to using these features for object recognition. The recognition proceeds by matching individual fea- tures to a database of features from known objects using a fast nearest-neighbor algorithm, followed by a Hough transform to identify clusters belonging to a sin- gle object, and finally performing verification through least-squares solution for consistent pose parameters. This approach to recognition can robustly identify objects among clutter and occlusion while achieving near real-time performance. Accepted for publication in the International Journal of Computer Vision, 2004. 1 1 Introduction Image matching is a fundamental aspect of many problems in computer vision, including object or scene recognition, solving for 3D structure from multiple images, stereo correspon- dence, and motion tracking. This paper describes image features that have many properties that make them suitable for matching differing images of an object or scene.
    [Show full text]
  • SIFTSIFT -- Thethe Scalescale Invariantinvariant Featurefeature Transformtransform
    SIFTSIFT -- TheThe ScaleScale InvariantInvariant FeatureFeature TransformTransform Distinctive image features from scale-invariant keypoints. David G. Lowe, International Journal of Computer Vision, 60, 2 (2004), pp. 91-110 Presented by Ofir Pele. Based upon slides from: - Sebastian Thrun and Jana Košecká - Neeraj Kumar Correspondence ■ Fundamental to many of the core vision problems – Recognition – Motion tracking – Multiview geometry ■ Local features are the key Images from: M. Brown and D. G. Lowe. Recognising Panoramas. In Proceedings of the the International Conference on Computer Vision (ICCV2003 ( Local Features: Detectors & Descriptors Detected Descriptors Interest Points/Regions <0 12 31 0 0 23 …> <5 0 0 11 37 15 …> <14 21 10 0 3 22 …> Ideal Interest Points/Regions ■ Lots of them ■ Repeatable ■ Representative orientation/scale ■ Fast to extract and match SIFT Overview Detector 1. Find Scale-Space Extrema 2. Keypoint Localization & Filtering – Improve keypoints and throw out bad ones 3. Orientation Assignment – Remove effects of rotation and scale 4. Create descriptor – Using histograms of orientations Descriptor SIFT Overview Detector 1.1. FindFind Scale-SpaceScale-Space ExtremaExtrema 2. Keypoint Localization & Filtering – Improve keypoints and throw out bad ones 3. Orientation Assignment – Remove effects of rotation and scale 4. Create descriptor – Using histograms of orientations Descriptor Scale Space ■ Need to find ‘characteristic scale’ for feature ■ Scale-Space: Continuous function of scale σ – Only reasonable kernel is
    [Show full text]
  • Lecture 12: Local Descriptors, SIFT & Single Object Recognition
    Lecture 12: Local Descriptors, SIFT & Single Object Recognition Professor Fei-Fei Li Stanford Vision Lab Fei-Fei Li Lecture 12 - 1 7-Nov-11 Midterm overview Mean: 70.24 +/- 15.50 Highest: 96 Lowest: 36.5 Fei-Fei Li Lecture 12 - 2 7-Nov-11 What we will learn today • Local descriptors – SIFT (Problem Set 3 (Q2)) – An assortment of other descriptors (Problem Set 3 (Q4)) • Recognition and matching with local features (Problem Set 3 (Q3)) – Matching objects with local features: David Lowe – Panorama stitching: Brown & Lowe – Applications Fei-Fei Li Lecture 12 - 3 7-Nov-11 Local Descriptors • We know how to detect points Last lecture (#11) • Next question: How to describe them for matching? ? Point descriptor should be: This lecture (#12) 1. Invariant 2. Distinctive Grauman Kristen credit: Slide Fei-Fei Li Lecture 12 - 4 7-Nov-11 Rotation Invariant Descriptors • Find local orientation – Dominant direction of gradient for the image patch • Rotate patch according to this angle – This puts the patches into a canonical orientation. Slide credit: Svetlana Lazebnik, Matthew Brown Matthew Lazebnik, Svetlana credit: Slide Fei-Fei Li Lecture 12 - 5 7-Nov-11 Orientation Normalization: Computation [Lowe, SIFT, 1999] • Compute orientation histogram • Select dominant orientation • Normalize: rotate to fixed orientation 0 2π Slide adapted from David Lowe David from adapted Slide Fei-Fei Li Lecture 12 - 6 7-Nov-11 The Need for Invariance • Up to now, we had invariance to – Translation – Scale – Rotation • Not sufficient to match regions under viewpoint changes – For this, we need also affine adaptation Slide credit: Tinne Tuytelaars Tinne credit: Slide Fei-Fei Li Lecture 12 - 7 7-Nov-11 Affine Adaptation • Problem: – Determine the characteristic shape of the region.
    [Show full text]