Fast Rigid Motion Segmentation Via Incrementally-Complex Local Models

Total Page:16

File Type:pdf, Size:1020Kb

Fast Rigid Motion Segmentation Via Incrementally-Complex Local Models Fast Rigid Motion Segmentation via Incrementally-Complex Local Models Fernando Flores-Mangas Allan D. Jepson Department of Computer Science, University of Toronto fmangas,[email protected] Abstract The problem of rigid motion segmentation of trajectory data under orthography has been long solved for non- degenerate motions in the absence of noise. But because real trajectory data often incorporates noise, outliers, mo- tion degeneracies and motion dependencies, recently pro- posed motion segmentation methods resort to non-trivial (a) Original video frames (b) Class-labeled trajectories representations to achieve state of the art segmentation ac- curacies, at the expense of a large computational cost. This paper proposes a method that dramatically reduces this cost (by two or three orders of magnitude) with minimal accu- racy loss (from 98.8% achieved by the state of the art, to 96.2% achieved by our method on the standard Hopkins 155 dataset). Computational efficiency comes from the use of a (c) Model support region (d) Model inliers and control points simple but powerful representation of motion that explicitly incorporates mechanisms to deal with noise, outliers and motion degeneracies. Subsets of motion models with the best balance between prediction accuracy and model com- plexity are chosen from a pool of candidates, which are then used for segmentation. (e) Model residuals (f) Segmentation result Figure 1: Model instantiation and segmentation. a) f th orig- 1. Rigid Motion Segmentation inal frame, Italian Grand Prix c 2012 Formula 1. b) Class- Rigid motion segmentation (MS) consists on separating labeled, trajectory data W (red, green, blue and black cor- regions, features, or trajectories from a video sequence into respond to chassis, helmet, background and outlier classes f spatio-temporally coherent subsets that correspond to in- respectively). c) Spatially-local support subset W^ for a dependent, rigidly-moving objects in the scene (Figure1.b candidate motion in blue. d) Candidate motion model in- f or1.f). The problem currently receives renewed attention, liers in red, control points (ci from Eq.3) in white. e) f partly because of the extensive amount of video sources and Residuals (ri from Eq. 11) color-coded with label data, the applications that benefit from MS to perform higher level radial coordinate is logarithmic. f) Segmentation result. computer vision tasks, but also because the state of the art is reaching functional maturity. methods [11,6,8] balance model complexity with modeling Motion Segmentation methods are widely diverse, but accuracy and have been successful at incorporating more most capture only a small subset of constraints or alge- of these aspects into a single formulation. For instance, braic properties from those that govern the image formation in [8] most model parameters are estimated automatically process of moving objects and their corresponding trajec- from the data, including the number of independent motions tories, such as the rank limit theorem [9, 10], the linear in- and their complexity, as well as the segmentation labels (in- dependence constraint (between trajectories from indepen- cluding outliers). However, because of the large number of dent motions) [2, 13], the epipolar constraint [7], and the necessary motion hypotheses that need to be instantiated, reduced rank property [11, 15, 13]. Model-selection based as well as the varying and potentially very large number of 1 model parameters that must be estimated, the flexibility of- 2D affine transformation models the motion of an arbitrary fered by this method comes at a large computational cost. plane, and the epipolar residuals account for relative depths. Current state of the art methods follow the trend of us- Crucially, these two components can be computed sepa- ing sparse low-dimensional subspaces to represent trajec- rately and incrementally, which enables an explicit mech- tory data. This representation is then fed into a clustering anism to deal with motion degeneracy. algorithm to obtain a segmentation result. A prime example In the context of 3D motion, a motion is degenerate when of this type of method is Sparse Subspace Clustering (SSC) the trajectories originate from a planar (or linear) object, or [3] in which each trajectory is represented as a sparse linear when neither the camera nor the imaged object exercise all combination of a few other basis trajectories. The assump- of their degrees of freedom, such as when the object only tion is that the basis trajectories must belong to the same translates, or when the camera only rotates. These are com- rigid motion as the reconstructed trajectory (or else, the re- mon situations in real world video sequences. The incre- construction would be impossible). When the assumption mental nature of the decompositions described above, fa- is true, the sparse mixing coefficients can be interpreted as cilitate the transition between degenerate motions and non- the connectivity weights of a graph (or a similarity matrix), degenerate ones. which is then (spectral) clustered to obtain a segmentation result. At the time of publication, SSC produced segmenta- Planar Model Under orthography, the projection of tra- tion results three times more accurate than the best prede- jectories from a planar surface can be modeled with the cessor. The practical downside, however, is the inherently affine transformation: large computational cost of finding the optimal sparse rep- 2 xc 3 2 xw 3 2 xw 3 resentation, which is at least cubic on the number of trajec- D t yc = yw = Aw!c yw ; (1) tories. 4 5 0> 1 4 5 2D 4 5 1 1 1 The work of [14] also falls within the class of subspace separation algorithms. Their approach is based on cluster- where D 2 R2×2 is an invertible matrix, and t 2 R2 is ing the principal angles (CPA) of the local subspaces associ- a translation vector. Trajectory coordinates (xw; yw) are in ated to each trajectory and its nearest neighbors. The clus- the plane’s reference frame (modulo a 2D affine transfor- tering re-weights a traditional metric of subspace affinity mation) and (xc; yc) are image coordinates. between principal angles. Re-weighted affinities are then Now, let W 2 R2F ×P be matrix of trajectory data that used for segmentation. The approach produces segmenta- contains the x and y image coordinates of P feature points tion results with accuracies similar to those of SSC, but the tracked through F frames, as in computational cost is close to 10 times bigger than SSC’s. 2 3 x1;1 ··· x1;P In this work we argue that competitive segmentation re- 6 y1;1 ··· y1;P 7 6 7 sults are possible using a simple but powerful representation 6 . .. 7 W = 6 . 7 : (2) of motion that explicitly incorporates mechanisms to deal 6 7 with noise, outliers and motion degeneracies. The proposed 4 xF;1 ··· xF;P 5 method is approximately 2 or 3 orders of magnitude faster yF;1 ··· yF;P than [3] and [14] respectively, currently considered the state To compute the parameters of A2D from trajectory data, let 2f×3 of the art. C = [c1; c2; c3] 2 R be three columns (three full tra- f 2f−1 2f > 1.1. Affine Motion jectories) from W, and let ci = [ci ; ci ] be the x and y coordinates of the i-th control trajectory at frame f. Then Projective Geometry is often used to model the image the transformation between points from an arbitrary source motion of trajectories from rigid objects between pairs of frame s to a target frame f can be written as: frames. However, alternative geometric relationships that cf cf cf cs cs cs facilitate parameter computation have also been proven use- 1 2 3 = As!f 1 2 3 ; (3) ful for this purpose. For instance, in perspective projection, 1 1 1 2D 1 1 1 general image motion from rigid objects can be modeled and As!f can be simply computed as: via the composition of two elements: a 2D homography, 2D −1 and parallax residual displacements [5]. The homography cf cf cf cs cs cs As!f = 1 2 3 1 2 3 : (4) describes the motion of an arbitrary plane, and the parallax 2D 1 1 1 1 1 1 residuals account for relative depths, that are unaccounted for by the planar surface model. The inverse in the right-hand side matrix of Eq.4 exists so s Under orthography, in contrast, image motion of rigid long as the points ci are not collinear. For simplicity we s!f f s objects can be modeled via the composition of a 2D affine refer to A2D as A2D and consequently A2D is the identity transformation plus epipolar residual displacements. The matrix. 3D Model In order to upgrade a planar (degenerate) that after this stage, the segmentation that results from the model into a full 3D one, relative depth must be accounted optimal model combination could be reported as a segmen- using the epipolar residual displacements. This means ex- tation result (x5). tending Eq.1 with a direction vector, scaled by the corre- The third stage incorporates the results from a set of sponding relative depth of each point, as in: model combinations that are closest to the optimal. Seg- mentation results are aggregated into an affinity matrix, 2 xc 3 2 xw 3 2 a 3 D ~t 13 which is then passed to a spectral clustering algorithm to yc = yw + δzw a : (5) 4 5 ~0> 1 4 5 4 23 5 produce the final segmentation result. This refinement stage 1 1 0 generally results in improved accuracy and reduced seg- mentation variability (x6). The depth δzw is relative to the arbitrary plane whose mo- tion is modeled by A2D; a point that lies on such plane would have δzw = 0.
Recommended publications
  • Pnp-Net: a Hybrid Perspective-N-Point Network
    PnP-Net: A hybrid Perspective-n-Point Network Roy Sheffer1, Ami Wiesel1;2 (1) The Hebrew University of Jerusalem (2) Google Research Abstract. We consider the robust Perspective-n-Point (PnP) problem using a hybrid approach that combines deep learning with model based algorithms. PnP is the problem of estimating the pose of a calibrated camera given a set of 3D points in the world and their corresponding 2D projections in the image. In its more challenging robust version, some of the correspondences may be mismatched and must be efficiently dis- carded. Classical solutions address PnP via iterative robust non-linear least squares method that exploit the problem's geometry but are ei- ther inaccurate or computationally intensive. In contrast, we propose to combine a deep learning initial phase followed by a model-based fine tuning phase. This hybrid approach, denoted by PnP-Net, succeeds in estimating the unknown pose parameters under correspondence errors and noise, with low and fixed computational complexity requirements. We demonstrate its advantages on both synthetic data and real world data. 1 Introduction Camera pose estimation is a fundamental problem that is used in a wide vari- ety of applications such as autonomous driving and augmented reality. The goal of perspective-n-point (PnP) is to estimate the pose parameters of a calibrated arXiv:2003.04626v1 [cs.CV] 10 Mar 2020 camera given the 3D coordinates of a set of objects in the world and their corre- sponding 2D coordinates in an image that was taken (see Fig. 1). Currently, PnP can be efficiently solved under both accurate and low noise conditions [15,17].
    [Show full text]
  • A Multi-Temporal Image Registration Method Based on Edge Matching and Maximum Likelihood Estimation Sample Consensus
    A MULTI-TEMPORAL IMAGE REGISTRATION METHOD BASED ON EDGE MATCHING AND MAXIMUM LIKELIHOOD ESTIMATION SAMPLE CONSENSUS Yan Songa, XiuXiao Yuana, Honggen Xua aSchool of Remote Sensing and Information Engineering School,Wuhan University, 129 Luoyu Road,Wuhan,China 430079– [email protected] Commission III, WG III/1 KEY WORDS: Change Detection, Remote Sensing, Registration, Edge Matching, Maximum Likelihood Sample Consensus ABSTRACT: In the paper, we propose a newly registration method for multi-temporal image registration. Multi-temporal image registration has two difficulties: one is how to design matching strategy to gain enough initial correspondence points. Because the wrong matching correspondence points are unavoidable, we are unable to know how many outliers, so the second difficult of registration is how to calculate true registration parameter from the initial point set correctly. In this paper, we present edge matching to resolve the first difficulty, and creatively introduce maximum likelihood estimation sample consensus to resolve the robustness of registration parameter calculation. The experiment shows, the feature matching we utilized has better performance than traditional normalization correlation coefficient. And the Maximum Likelihood Estimation Sample Conesus is able to solve the true registration parameter robustly. And it can relieve us from defining threshold. In experiment, we select a pair of IKONOS imagery. The feature matching combined with the Maximum Likelihood Estimation Sample Consensus has robust and satisfying registration result. 1. INTRODUCTION majority. So in the second stage, we pay attention on how to calculate registration parameter robustly. These will make us Remote sensing imagery change detection has gain more and robustly calculate registration parameter successfully, even more attention in academic and research society in near decades.
    [Show full text]
  • Density Adaptive Point Set Registration
    Density Adaptive Point Set Registration Felix Jaremo¨ Lawin, Martin Danelljan, Fahad Shahbaz Khan, Per-Erik Forssen,´ Michael Felsberg Computer Vision Laboratory, Department of Electrical Engineering, Linkoping¨ University, Sweden {felix.jaremo-lawin, martin.danelljan, fahad.khan, per-erik.forssen, michael.felsberg}@liu.se Abstract Probabilistic methods for point set registration have demonstrated competitive results in recent years. These techniques estimate a probability distribution model of the point clouds. While such a representation has shown promise, it is highly sensitive to variations in the den- sity of 3D points. This fundamental problem is primar- ily caused by changes in the sensor location across point sets. We revisit the foundations of the probabilistic regis- tration paradigm. Contrary to previous works, we model the underlying structure of the scene as a latent probabil- ity distribution, and thereby induce invariance to point set density changes. Both the probabilistic model of the scene and the registration parameters are inferred by minimiz- ing the Kullback-Leibler divergence in an Expectation Max- imization based framework. Our density-adaptive regis- tration successfully handles severe density variations com- monly encountered in terrestrial Lidar applications. We perform extensive experiments on several challenging real- world Lidar datasets. The results demonstrate that our ap- proach outperforms state-of-the-art probabilistic methods for multi-view registration, without the need of re-sampling. Figure 1. Two example Lidar scans (top row), with signifi- cantly varying density of 3D-points. State-of-the-art probabilistic method [6] (middle left) only aligns the regions with high density. This is caused by the emphasis on dense regions, as visualized by 1.
    [Show full text]
  • Accurate Motion Estimation Through Random Sample Aggregated Consensus
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. X, NO. X, AUGUST 2016 1 Accurate Motion Estimation through Random Sample Aggregated Consensus Martin Rais, Gabriele Facciolo, Enric Meinhardt-Llopis, Jean-Michel Morel, Antoni Buades, and Bartomeu Coll Abstract—We reconsider the classic problem of estimating accurately a 2D transformation from point matches between images containing outliers. RANSAC discriminates outliers by randomly generating minimalistic sampled hypotheses and verifying their consensus over the input data. Its response is based on the single hypothesis that obtained the largest inlier support. In this article we show that the resulting accuracy can be improved by aggregating all generated hypotheses. This yields RANSAAC, a framework that improves systematically over RANSAC and its state-of-the-art variants by statistically aggregating hypotheses. To this end, we introduce a simple strategy that allows to rapidly average 2D transformations, leading to an almost negligible extra computational cost. We give practical applications on projective transforms and homography+distortion models and demonstrate a significant performance gain in both cases. Index Terms—RANSAC, correspondence problem, robust estimation, weighted aggregation. F 1 INTRODUCTION EVERAL applications in computer vision such as im- the parameters of the underlying transformation by regres- S age alignment, panoramic mosaics, 3D reconstruction, sion. RANSAC achieves good accuracy when observations motion tracking, object recognition, among others, rely on are noiseless, even with a significant number of outliers in feature-based approaches. These approaches can be divided the input data. in three distinct steps. First, characteristic image features are Instead of using every sample in the dataset to per- detected on the input images.
    [Show full text]
  • Automatic Panoramic Image Stitching Using Invariant Features
    Automatic Panoramic Image Stitching using Invariant Features Matthew Brown and David G. Lowe {mbrown|lowe}@cs.ubc.ca Department of Computer Science, University of British Columbia, Vancouver, Canada. Abstract can provide very accurate registration, but they require a close initialisation. Feature based registration does not re- This paper concerns the problem of fully automated quire initialisation, but traditional feature matching meth- panoramic image stitching. Though the 1D problem (single ods (e.g., correlation of image patches around Harris cor- axis of rotation) is well studied, 2D or multi-row stitching is ners [Har92, ST94]) lack the invariance properties needed more difficult. Previous approaches have used human input to enable reliable matching of arbitrary panoramic image or restrictions on the image sequence in order to establish sequences. matching images. In this work, we formulate stitching as a In this paper we describe an invariant feature based ap- multi-image matching problem, and use invariant local fea- proach to fully automatic panoramic image stitching. This tures to find matches between all of the images. Because of has several advantages over previous approaches. Firstly, this our method is insensitive to the ordering, orientation, our use of invariant features enables reliable matching of scale and illumination of the input images. It is also insen- panoramic image sequences despite rotation, zoom and illu- sitive to noise images that are not part of a panorama, and mination change in the input images. Secondly, by viewing can recognise multiple panoramas in an unordered image image stitching as a multi-image matching problem, we can dataset. In addition to providing more detail, this paper ex- automatically discover the matching relationships between tends our previous work in the area [BL03] by introducing the images, and recognise panoramas in unordered datasets.
    [Show full text]
  • Tracking Algorithms for Tpcs Using Consensus-Based Robust Estimators
    Tracking Algorithms for TPCs using Consensus-Based Robust Estimators J. C. Zamora∗, G. F. Fortino Instituto de Fisica, Universidade de Sao Paulo, Sao Paulo 05508-090, Brazil Abstract A tracking algorithm1 based on consensus-robust estimators was implemented for the analysis of experiments with time-projection chambers. In this work, few algorithms beyond RANSAC were successfully tested using experimental data taken with the AT-TPC, ACTAR and TexAT detectors. The present track- ing algorithm has a better inlier-outlier detection than the simple sequential RANSAC routine. Modifications in the random sampling and clustering were included to improve the tracking efficiency. Very good results were obtained in all the test cases, in particular for fitting short tracks in the detection limit. Keywords: Time Projection Chambers, Active Target, Tracking Algorithm, RANSAC, clustering 1. Introduction Time Projection Chambers (TPC) are among the most efficient devices exist- ing for charged-particle tracking. They enable measurements of energy loss and position hits in the 3-dimensional space which are employed for particle identifi- cation and momentum reconstruction. TPCs have been successfully used during arXiv:2011.07138v1 [physics.ins-det] 13 Nov 2020 the last three decades as tracking systems in many particle-physics experiments, e.g., TOPAZ [1], STAR [2] or ALICE [3]. ∗Corresponding author Email address: [email protected] (J. C. Zamora ) 1Code available at https://github.com/jczamorac/Tracking_RANSAC.git Preprint submitted to NIM A November 17, 2020 In the past few years, the operation of TPCs in Active Target (AT) mode have gained great attention for investigation in nuclear physics. The AT prin- ciple brings the advantage to use a gas as detector material and target medium simultaneously.
    [Show full text]
  • Two-View Geometry Estimation by Random Sample and Consensus
    Two-View Geometry Estimation by Random Sample and Consensus A Dissertation Presented to the Faculty of the Electrical Engineering of the Czech Technical University in Prague in Partial Fulfillment of the Re- quirements for the Ph.D. Degree in Study Programme No. P2612 - Electrotech- nics and Informatics, branch No. 3902V035 - Artificial Intelligence and Bio- cybernetics, by Ondrejˇ Chum July 1, 2005 Thesis Advisor Dr. Ing. Jirˇ´ı Matas Center for Machine Perception Department of Cybernetics Faculty of Electrical Engineering Czech Technical University in Prague Karlovo nam´ estˇ ´ı 13, 121 35 Prague 2, Czech Republic fax: +420 224 357 385, phone: +420 224 357 465 http://cmp.felk.cvut.cz Abstract The problem of model parameters estimation from data with a presence of outlier measurements often arises in computer vision and methods of robust estimation have to be used. The RANSAC algorithm introduced by Fishler and Bolles in 1981 is the a widely used robust estimator in the field of computer vision. The algorithm is capable of providing good estimates from data contaminated by large (even significantly more than 50%) fraction of outliers. RANSAC is an optimization method that uses a data-driven random sampling of the parameter space to find the extremum of the cost function. Samples of data define points of the parameter space in which the cost function is evaluated and model parameters with the best score are output. This thesis provides a detailed analysis of RANSAC, which is recast as time-constrained op- timization – a solution that is optimal with certain confidence is sought in the shortest possible time.
    [Show full text]
  • Comparative the Performance of Robust Linear Regression Methods 817
    Applied Mathematical Sciences, Vol. 13, 2019, no. 17, 815 - 822 HIKARI Ltd, www.m-hikari.com https://doi.org/10.12988/ams.2019.97106 Comparative the Performance of Robust Linear Regression Methods and its Application Unchalee Tonggumneada and Nikorn Saengngamb a Department of Mathematics and Computer Science, Faculty of Science and Technology, Rajamangala University of Technology Thanyaburi, Phatum Thani 12110, Thailand b Department of Technical Education , Faculty of Technical Education, Rajamangala University of Technology Thanyaburi, Phatum Thani 12110, Thailand This article is distributed under the Creative Commons by-nc-nd Attribution License. Copyright © 2019 Hikari Ltd. Abstract This research aims to apply the robust estimation method with actual data. M- estimation, quantile regression and Random Sample Consensus (RANSAC) are considered. The result show that, when no outlier, the parameter estimate using M- estimation, quantile regression and random sample consensus (RANSAC) produce a very similary. When the rate of outliers increasing is also conducted the parameter estimate using RANSAC method more different from M- estimation and quantile regression. The RANSAC method is more robust to outlier than the M- estimation and quantile regression method. Which can be clearly seen that, The RANSAC methods are more robust against the outlier than M- estimation and quantile regression ,because in parameter estimation do not include the outlier into the model, only inliers are used to run the model. Keywords: robust regression, M- estimation, quantile regression, Random Sample Consensus. 1 Introduction Regression analysis is the common and useful statistical tool which can be used to quantify the relationship between a response variable (y) and explanatory variables (x), Ordinary Least Square (OLS) is A common used for parameter estimate.
    [Show full text]
  • Automated Fast Initial Guess in Digital Image Correlation
    An International Journal for Experimental Mechanics Automated Fast Initial Guess in Digital Image Correlation † ‡ Z. Wang*,M.Vo, H. Kieu* and T. Pan *Department of Mechanical Engineering, The Catholic University of America, Washington, DC 20064, USA † Robotics Institute, Carnegie Mellon University, Pittsburgh, PA 15213, USA ‡ Department of Civil Engineering, The Catholic University of America, Washington, DC 20064, USA ABSTRACT: A challenging task that has hampered the fully automatic processing of the digital image correlation (DIC) technique is the initial guess when large deformation and rotation are present. In this paper, a robust scheme combining the concepts of a scale-invariant feature transform (SIFT) algorithm and an improved random sample consensus (iRANSAC) algorithm is employed to conduct an automated fast initial guess for the DIC technique. The scale-invariant feature transform algorithm can detect a certain number of matching points from two images even though the corresponding deformation and rotation are large or the images have periodic and identical patterns. After removing the wrong matches with the improved random sample consensus algorithm, the three pairs of closest and non-collinear matching points serve for the purpose of initial guess calculation. The validity of the technique is demonstrated by both computer simulation and real experiment. KEY WORDS: digital image correlation, improved random sample consensus, initial guess, scale-invariant feature transform Introduction analysis is to minimise the coefficient C in Equation (1) to find the best matching. For a representative Digital image correlation (DIC) technique is one of the pixel (x ,y ) to be analysed in the reference image, a most widely used methods for shape, motion and 0 0 square pattern of N =(2M +1)×(2M +1)pixelswithits deformation measurements [1].
    [Show full text]
  • Robust Regression Sample Consensus Estimator Model: Performed on Inliers
    International Journal of Engineering Applied Sciences and Technology, 2018 Vol. 3, Issue 6, ISSN No. 2455-2143, Pages 22-28 Published Online October 2018 in IJEAST (http://www.ijeast.com) ROBUST REGRESSION SAMPLE CONSENSUS ESTIMATOR MODEL: PERFORMED ON INLIERS R.Akila Dr.J.Ravi S.Dickson Assistant Professor Associate Professor Assistant Professor Dept. of Maths & Stat. Dept. of Maths & Stat. Dept. of Maths & Stat. Vivekanandha College for Women Vivekanandha College for Women Vivekanandha College for Woemn Tiruchengode, Namakkal Tiruchengode, Namakkal Tiruchengode, Namakkal Tamilnadu, India Tamilnadu, India Tamilnadu, India Abstract-Robust regression play vital role in many fields Lindquist (1953); and Glass, Peckham, and Sanders (1972). specially image processing. Corner detection is the one of the Countless studies have shown that even when classic important role in image process. Much corner detection is to parametric tests are robust to Type I errors, they are usually develop day to day; such corner detection is based on robust considerably less powerful than their modern robust regression. SAmple Consensus (SAC) has been popular in counterparts [2]. The RANSAC families formulate regression regression problems with samples contaminated with with outliers as a minimization problem. This formulation is outliers. It has been a milestone of many researches on similar with least square method, which minimize sum of robust estimators, but there are a few survey and squared error values. Robust fitting algorithms are mostly performance analysis on them. According to SAC family to based on sampling strategies, hypotheses are generated and analyze the performance evaluation based on being accurate their support is measured in the point cloud.
    [Show full text]
  • Recursive RANSAC: Multiple Signal Estimation with Outliers
    Recursive RANSAC: Multiple Signal Estimation with Outliers Peter C. Niedfeldt, Randal W. Beard Brigham Young University, Provo, UT USA ([email protected], [email protected]). Abstract: The random sample consensus (RANSAC) algorithm is frequently used in computer vision to estimate the parameters of a signal in the presence of noisy and even spurious observations called gross errors. Instead of just one signal, we desire to estimate the parameters of multiple signals, where at each time step a set of observations of generated from the underlying signals and gross errors are received. In this paper, we develop the recursive RANSAC (RRANSAC) algorithm to solve the inherent data association problem and recursively estimate the parameters of multiple signals without prior knowledge of the number of true signals. We compare the performance of RRANSAC with several existing algorithms, and also demonstrate the capabilities of RRANSAC in an aerial geolocation problem. 1. INTRODUCTION they require solving a nonlinear minimization problem. See Rousseeuw and Leroy [1987], Huber [1996]. The random sample consensus (RANSAC) algorithm de- Regression analysis describes a collection of techniques veloped by Fischler and Bolles [1981] is a novel approach used to determine the relationships between inputs to a to regression analysis. Many previous algorithms formed system and the outputs. A common scenario is where a model using all or most of the available data set, and the system model is assumed to be known while the then removed observations inconsistent with the model specific parameters of the underlying signals are to be before producing a final estimate. Alternatively, RANSAC estimated. One well known algorithm is least-squares (LS), approaches the problem by forming numerous model es- which minimizes the mean-squared error between a set timates using the minimum number of required points of observations and the estimated model.
    [Show full text]
  • Feature Matching with Improved SIRB Using RANSAC by Meet Palod, Manas Joshi, Amber Jain, Viroang Rawat & Prof
    Global Journal of Computer Science and Technology: F Graphics & Vision Volume 21 Issue 1 Version 1.0 Year 2021 Type: Double Blind Peer Reviewed International Research Journal Publisher: Global Journals Online ISSN: 0975-4172 & Print ISSN: 0975-4350 Feature Matching with Improved SIRB using RANSAC By Meet Palod, Manas Joshi, Amber Jain, Viroang Rawat & Prof. K.K. Sharma Abstract- In this paper, we suggest to improve the SIRB (SIFT (Scale-Invariant Feature Transform) and ORB (Oriented FAST and Rotated BRIEF)) algorithm by incorporating RANSAC to enhance the matching performance. We use multi-scale space to extract the features which are impervious to scale, rotation, and affine variations. Then the SIFT algorithm generates feature points and passes the interest points to the ORB algorithm. The ORB algorithm generates an ORB descriptor where Hamming distance matches the feature points. We propose to use RANSAC (Random Sample Consensus) to cut down on both the inliers in the form of noise and outliers drastically, to cut down on the computational time taken by the algorithm. This post- processing step removes redundant key points and noises. This computationally effective and accurate algorithm can also be used in handheld devices where their limited GPU acceleration is not able to compensate for the computationally expensive algorithms like SIFT and SURF. Experimental results advocate that the proposed algorithm achieves good matching, improves efficiency, and makes the feature point matching more accurate with scale in-variance taken into consideration. GJCST-F Classification: G.4 FeatureMatchingwithImprovedSIRBusingRANSAC Strictly as per the compliance and regulations of: © 2021. Meet Palod, Manas Joshi, Amber Jain, Viroang Rawat & Prof.
    [Show full text]