Automatic Generation of a 3D City Model

Total Page:16

File Type:pdf, Size:1020Kb

Automatic Generation of a 3D City Model UNIVERSITY OF CASTILLA-LA MANCHA ESCUELA SUPERIOR DE INFORMÁTICA COMPUTER ENGINEERING DEGREE DEGREE FINAL PROJECT Automatic generation of a 3D city model David Murcia Pacheco June, 2017 AUTOMATIC GENERATION OF A 3D CITY MODEL Escuela Superior de Informática UNIVERSITY OF CASTILLA-LA MANCHA ESCUELA SUPERIOR DE INFORMÁTICA Information Technology and Systems SPECIFIC TECHNOLOGY OF COMPUTER ENGINEERING DEGREE FINAL PROJECT Automatic generation of a 3D city model Author: David Murcia Pacheco Director: Dr. Félix Jesús Villanueva Molina June, 2017 David Murcia Pacheco Ciudad Real – Spain E-mail: [email protected] Phone No.:+34 625 922 076 c 2017 David Murcia Pacheco Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.3 or any later version published by the Free Software Foundation; with no Invariant Sections, no Front-Cover Texts, and no Back-Cover Texts. A copy of the license is included in the section entitled "GNU Free Documentation License". i TRIBUNAL: Presidente: Vocal: Secretario: FECHA DE DEFENSA: CALIFICACIÓN: PRESIDENTE VOCAL SECRETARIO Fdo.: Fdo.: Fdo.: ii Abstract HIS document collects all information related to the Degree Final Project (DFP) of Com- T puter Engineering Degree of the student David Murcia Pacheco, tutorized by Dr. Félix Jesús Villanueva Molina. This work has been developed during 2016 and 2017 in the Escuela Superior de Informática (ESI), in Ciudad Real, Spain. It is based in one of the proposed sub- jects by the faculty of this university for this year, called "Generación automática del modelo en 3D de una ciudad". This work is part of a bigger project that the ESI is working on, which final goal is to create an operative Smart City model of the city. This bigger project is composed of different works like this (as it is too ambitious to be carried out by a single person). The project subject of this paper is focused on supporting the 3D modeling of the city. Data for doing this work is available and free in the Internet, but it needs to be processed and adjusted so it can be standardized to be used. Main focus of this DFP is to study a quite new technology that is (and is going to be even more) useful and used in the future. This technology is Light Detection and Rang- ing (LIDAR). Other technologies will be studied in parallel. It can be said that LIDAR data is a cloud of points resulting from a scanning process. Lots of devices and software is being de- veloped for this technology, as industry and research is taking advantage of its possibilities. Big scale Airborne LIDAR technology is useful for purposes such as agriculture, forest plan- ning, flood models, mapping, mining, archeology... There is also small scale LIDAR, used for scanning smaller objects or spaces. In this case, this technology is useful for 3D scanning (creating models for, e.g., 3D printers), architecture, 3D gaming, CAD modeling... While working on this project, it has been seen that the DFP is making use of a technology that is going to be very important in the medium and far future. iii Thanks To faculty of ESI. The quality and dedication of their work is outstanding. To all my Vestas colleagues, with whom i try to improve every day, personal and profes- sionally. To my elders brothers Emilio, Pedro Jose and Alberto. I’ve learned so much from them... To my wife Eva, for her support and understanding specially during my study years. To Carla, my new born and first daughter, for helping me value dedication needed to raise another person. This is one of the biggest sacrifice that, as humans, we can choose to make. Sacrifice which some year ago my parents Emilio and Ramona did for me. My biggest thanks is for them. David iv To my wife, Eva v Contents Abstract iii Thanks iv Contents vi List of Tables vii List of Figures viii List of Listings ix LIST OF ACRONYMS x 1 INTRODUCTION 1 1.1 Document Structure . .3 2 GOALS OF DFP 4 2.1 MAIN GOAL . .4 2.2 SPECIFIC GOALS AND DEFINED REQUIREMENTS . .4 2.3 CHALLENGES, LIMITATIONS AND SCOPE . .5 2.4 RESOURCES USED . .6 2.4.1 Hardware Resources . .6 2.4.2 Software Resources . .6 3 ANTECEDENTS AND STATE OF ART 8 3.1 LIDAR . .8 3.1.1 Digital surface models and methods . .9 3.1.1.1 Data pre-preprocessing . .9 3.1.1.2 Ground Point Filtering . 10 3.1.1.3 Interpolation . 10 vi 3.1.2 DSM generation methods for different categories . 10 3.1.2.1 Surface-Based Adjustment . 11 3.1.2.2 Morphology-based filtering . 11 3.1.2.3 Triangulated Irregular Network (TIN)........... 12 3.1.3 Futures challenges for DSM generation . 12 3.1.4 LIDAR .las format Standard 1.4 . 12 3.1.4.1 LIDAR data types . 13 3.1.4.2 LIDAR header format . 13 3.1.4.3 LIDAR point data record format . 17 3.2 3D CITY RENDERING METHODS . 19 3.2.1 Building rendering from LIDAR data . 20 3.2.1.1 Identify blocks and buildings . 20 3.2.1.2 Roof modeling . 23 3.2.2 Building rendering merging aerial and ground views . 25 3.2.2.1 DSM processing . 25 3.2.2.2 Ground based model registration . 26 3.2.2.3 Model merging . 28 3.3 OSM . 29 3.3.1 How do OSM works . 30 3.3.1.1 OSM Geodata . 30 3.3.1.2 Accessing OSM data . 31 3.3.2 OSM XML data and file format . 32 3.4 .PLY FILE FORMAT . 33 4 METHOD 38 4.1 SOURCE CODE FILES AND CLASSES . 38 4.2 PROGRAM INPUT AND OUTPUT . 39 4.3 PROCESS FLOW DIAGRAM . 40 4.4 PROCESS FLOW AND CODE EXPLANATION . 41 5 RESULTS 52 5.1 BASIC TESTING . 52 5.1.1 Further testing . 54 5.2 OUTPUT FILES . 55 5.2.1 LIDAR output . 55 5.2.2 Modified .osm output . 58 vii 5.2.3 .ply output . 60 5.3 SCALABILITY TEST . 64 6 CONCLUSIONS AND PROPOSALS 66 6.1 CONCLUSIONS . 66 6.2 SATISFIED GOALS . 67 6.3 PROPOSALS AND FURTHER WORK . 67 A PREREQUISITES AND DEPENDENCES FOR EXECUTION 70 A.1 INSTALL NEEDED PACKAGES AND LIBRARIES . 70 A.2 DOCUMENTATION GENERATION WITH TEXMAKER . 72 A.3 PROBLEMS DURING INSTALLATION . 72 viii List of Tables 3.1 LIDAR v1.4 header . 14 3.2 LIDAR point data 10 format . 17 3.3 LIDAR point classification numeric assignation . 19 5.1 Scalability tests inputs and results . 64 ix List of Figures 1.1 Smart City Concept . .1 1.2 3D city model . .2 3.1 Airborne LIDAR scanning . .9 3.2 DSM image from New York Central Park . 11 3.3 3D detailed city render . 20 3.4 Delaunay triangulation in plane with circumcircles . 21 3.5 Centroidal Voronoi diagram . 22 3.6 Building walls definned by split Voronoi Cells . 23 3.7 Normal extracted from roof points . 24 3.8 Textured city model example . 25 3.9 Google Street View Car with scanners used for shape mapping . 26 3.10 Surveying vehicle principle . 27 3.11 Result from Building rendering merging aerial and ground views method . 28 3.12 Open Street Maps (OSM) view for Amsterdam city center . 29 3.13 Surveying device and software example . 30 3.14 Data layer for Ciudad Real on OSM web page . 32 3.15 Cow representation from a .ply file . 35 4.1 Process execution flow diagram . 40 5.1 LIDAR campus building points on CloudCompare . 56 5.2 LIDAR campus building and ground points on CloudCompare . 56 5.3 LIDAR ESI building and ground points on CloudCompare . 57 5.4 Entire city of Ciudad Real buildings and ground points on CloudCompare . 58 5.5 .osm representation of the UCLM campus on Blender . 60 5.6 ESI .ply roof on Blender . 61 5.7 Hospital General de Ciudad Real .ply roof on Blender . 61 5.8 LIDAR view of Alcorcón . 65 x List of Listings 3.1 OSM fragment on which the ESI is defined . 34 3.2 .ply header format . 36 3.3 .ply format for vertices and faces . 36 4.1 Input command check in «buildings_capture.cpp» . 41 4.2 .osm buffers, readers and writers initializing in «buildings_capture.cpp» . 41 4.3 .osm header define geographical box used in «buildings_capture.cpp» . 41 4.4 Reading lidar data. Method in «lidar.cpp» . 42 4.5 Set ready files, readers and writers for lidar files in «lidar.cpp» . 42 4.6 LIDAR Point selection and filtering in «lidar.cpp» . 42 4.7 Update header of LIDAR files and close them in «lidar.cpp» . 43 4.8 Statistics applied to find and remove height outliers in «lidar.cpp» . 43 4.9 Print work coordinates on command line in «buildings_capture.cpp» . 44 4.10 Start of .osm file processing in «buildings_capture.cpp» . 44 4.11 Processing of OSM building entities, contained in ways, in «osmium_han- dlers.cpp» . 45 4.12 Call of calculate_height() and create_roof_meshes() methods, in «buildings_- capture.cpp» . 46 4.13 Method for calculation of height of buildings, in «lidar.cpp» . 47 4.14 Method to create the point cloud model for roof tops, in «roof_mesh_gener- ator.cpp» . 49 4.15 Call writerhandler to modify data, and write .osm file, in «roof_mesh_gen- erator.cpp» . 50 4.16 Write of the height tag on buildings for new .osm file, in «osmium_han- dlers.cpp» . 50 4.17 Program finalization and printing result, in«roof_mesh_generator.cpp» .
Recommended publications
  • Benchmarks on WWW Performance
    The Scalability of X3D4 PointProperties: Benchmarks on WWW Performance Yanshen Sun Thesis submitted to the Faculty of the Virginia Polytechnic Institute and State University in partial fulfillment of the requirements for the degree of Master of Science in Computer Science and Application Nicholas F. Polys, Chair Doug A. Bowman Peter Sforza Aug 14, 2020 Blacksburg, Virginia Keywords: Point Cloud, WebGL, X3DOM, x3d Copyright 2020, Yanshen Sun The Scalability of X3D4 PointProperties: Benchmarks on WWW Performance Yanshen Sun (ABSTRACT) With the development of remote sensing devices, it becomes more and more convenient for individual researchers to acquire high-resolution point cloud data by themselves. There have been plenty of online tools for the researchers to exhibit their work. However, the drawback of existing tools is that they are not flexible enough for the users to create 3D scenes of a mixture of point-based and triangle-based models. X3DOM is a WebGL-based library built on Extensible 3D (X3D) standard, which enables users to create 3D scenes with only a little computer graphics knowledge. Before X3D 4.0 Specification, little attention has been paid to point cloud rendering in X3DOM. PointProperties, an appearance node newly added in X3D 4.0, provides point size attenuation and texture-color mixing effects to point geometries. In this work, we propose an X3DOM implementation of PointProperties. This implementation fulfills not only the features specified in X3D 4.0 documentation, but other shading effects comparable to the effects of triangle-based geometries in X3DOM, as well as other state-of-the-art point cloud visualization tools.
    [Show full text]
  • Cloudcompare Point Cloud Processing Workshop Daniel Girardeau-Montaut
    CloudCompare Point Cloud Processing Workshop Daniel Girardeau-Montaut www.cloudcompare.org @CloudCompareGPL PCP 2019 [email protected] December 4-5, 2019, Stuttgart, Germany Outline About the project Quick overview of the software capabilities Point cloud processing with CloudCompare The project 2003: PhD for EDF R&D EDF Main French power utility More than 150 000 employees worldwide 2 000 @ R&D (< 2%) 200 know about CloudCompare (< 0.2%) Sales >75 B€ > 200 dams + 58 nuclear reactors (19 plants) EDF and Laser Scanning EDF = former owner of Mensi (now Trimble Laser Scanning) Main scanning activity: as-built documentation Scanning a single nuclear reactor building 2002: 3 days, 50 M. points 2014: 1.5 days, 50 Bn points (+ high res. photos) EDF and Laser Scanning Other scanning activities: Building monitoring (dams, cooling towers, etc.) Landslide monitoring Hydrology Historical preservation (EDF Foundation) PhD Change detection on 3D geometric data Application to Emergency Mapping Inspired by 9/11 post-attacks recovery efforts (see “Mapping Ground Zero” by J. Kern, Optech, Nov. 2001) TLS was used for: visualization, optimal crane placement, measurements, monitoring the subsidence of the wreckage pile, slurry wall monitoring, etc. CloudCompare V1 2004-2006 Aim: quickly detecting changes by comparing TLS point clouds… with a CAD mesh or with another (high density) cloud CloudCompare V2 2007: “Industrialization” of CloudCompare … for internal use only! Rationale: idle reactor = 6 M€ / day acquired data
    [Show full text]
  • Seamless Texture Mapping of 3D Point Clouds
    Seamless Texture Mapping of 3D Point Clouds Dan Goldberg Mentor: Carl Salvaggio Chester F. Carlson Center for Imaging Science, Rochester Institute of Technology Rochester, NY November 25, 2014 Abstract The two similar, quickly growing fields of computer vision and computer graphics give users the ability to immerse themselves in a realistic computer generated environment by combining the ability create a 3D scene from images and the texture mapping process of computer graphics. The output of a popular computer vision algorithm, structure from motion (obtain a 3D point cloud from images) is incomplete from a computer graphics standpoint. The final product should be a textured mesh. The goal of this project is to make the most aesthetically pleasing output scene. In order to achieve this, auxiliary information from the structure from motion process was used to texture map a meshed 3D structure. 1 Introduction The overall goal of this project is to create a textured 3D computer model from images of an object or scene. This problem combines two different yet similar areas of study. Computer graphics and computer vision are two quickly growing fields that take advantage of the ever-expanding abilities of our computer hardware. Computer vision focuses on a computer capturing and understanding the world. Computer graphics con- centrates on accurately representing and displaying scenes to a human user. In the computer vision field, constructing three-dimensional (3D) data sets from images is becoming more common. Microsoft's Photo- synth (Snavely et al., 2006) is one application which brought attention to the 3D scene reconstruction field. Many structure from motion algorithms are being applied to data sets of images in order to obtain a 3D point cloud (Koenderink and van Doorn, 1991; Mohr et al., 1993; Snavely et al., 2006; Crandall et al., 2011; Weng et al., 2012; Yu and Gallup, 2014; Agisoft, 2014).
    [Show full text]
  • Oracle Spatial 11G Release 2 New Features Deep Dive Siva Ravada Director of Development, Oracle Spatial
    Title of Presentation Oracle Spatial 11g Release 2 New Features Deep Dive Siva Ravada Director of Development, Oracle Spatial Michael Smith US Army Corps of Engineers Agenda • Introduction to Oracle Spatial <Insert Picture Here> • Spatial Engine • SQL Developer • GeoRaster • Network Data Model • Moving Objects • 3D Support Oracle Spatial Features – Oracle Locator: Feature of Oracle Database XE, SE, EE – Oracle Spatial: Priced option to HTTP Oracle Database EE Fusion Middleware – MapViewer*: Java application and map rendering feature of Fusion MapViewer Middleware – Bundled Map Content: Major roads, JDBC administrative boundaries (city, county, state, country) - worldwide Oracle Database coverage from Navteq Oracle Locator Oracle Spatial Bundled Map Content Supports all Geospatial Data types Polygons (admin, sales territories, Locations high risk zones) Networks (points of interest) (roads, utilities) Data Imagery 3D data type (satellite imagery) Topology (city models) (data provider) LIDAR Data Type TIN Data Type Open Platform for Leading Applications and Tools MapInfo Leica ADE Bentley Manifold Oracle Spatial 11g Enabled Applications Scrollable, Interactive Maps Spatial Web Services 3D, Point Clouds, and LIDAR Geocoding & Routing Oracle BI Dashboards Raster Imagery Spatial Engine Spatial on Exadata • We achieve performance improvements based on the exploitation of the processing power, bandwidth, and parallelism of the Exadata machine • Parallel query and partitioning help improve Spatial performance • All the row level functions are also parallel enabled to take advantage of parallel query • Speedups of up to 40 times (compared to a 1-core CPU box) for the typical SDO_ANYINTERACT spatial query • Spatial predicates are not currently pushed into the Exadata storage nodes 9 Spatial Index Parallel Creation on Exadata • Build a Local Spatial Index on Partitioned Table: each slave can build a different index partition in parallel.
    [Show full text]
  • TUESDAY MORNING Open Open Open
    TUESDAY MORNING Open Open Open. The Rise of Open Scientific Publishing and the 8:30 AM ­ Keynote Address Dr. Ken Kvamme Archaeological Discipline: Managing the Paradigm Shift SCE Auditorium Education and Dissemination 9:30 AM Coffee Break Room: SCE 216 Organizers: Arianna Traviglia, Xavier Rodier General Papers 9:50 AM Rocks­Macqueen, Doug, Tom Levy, The Last of the Publishers General ­ GIS 10:10 AM Roued­Cunliffe, Henriette, Peter Room: SCE House Jensen, What is Open Data in an Archaeological Context? 9:50 AM Caswell, Edward, An Analysis of 10:30 AM Brin, Adam, Francis Pierce­ Selected Variables that Affect the McManamon, Leigh Anne Ellison, The Production of Cost Surfaces Challenges of Open Access for 10:10 AM Becker, Daniel, Christian Willmes, Archaeological Repositories Andreas Bolten, María de Andrés­ 10:50 AM Kondo, Yasuhisa, Atsushi Noguchi, Herrero, Gerd­Christian Weniger, Best Practices and Challenges in Georg Bareth, Influence of Differences Promoting Open Science in in Digital Elevation Models and their Archaeology: Two Narratives from Resolution on Slope­Based Site Japan Catchment Modelling 11:10 AM Rodier, Xavier, Olivier Marlet, How to 10:30 AM Beusing, Ruth, Temporal­GIS Tools for Combine Archaeological Archives, Mixed and Meso­Scale Archaeological Linked Open Data and Academic Data Publication, a French Experience 10:50 AM James, Vivian, Context as Theory: 11:30 AM Buard, Pierre­Yves, Elisabeth Zadora­ Towards Unification of Computer Rio, Publishing an Archaeological Applications and Quantitative Methods Excavation Report in a Logicist in Archaeology Workflow 11:10 AM Sharp, Kayeleigh, Testing the ‘Small­ 11:50 AM Kansa, Eric, Anne Austin, Ixchel Site’ Approach with Multivariate Spatial Faniel, Sarah Kansa, Ran Boynter, Statistical and Archaeological Network Jennifer Jacobs, Phoebe France, A Analysis Reality Check: The Impact of Open Data in Archaeology 11:30 AM Groen, Mike, Geospatial Analysis, Predictive Modelling and Modern Clandestine Burials Same as it ever was..
    [Show full text]
  • Simulated Imagery Rendering Workflow for UAS-Based
    remote sensing Article Simulated Imagery Rendering Workflow for UAS-Based Photogrammetric 3D Reconstruction Accuracy Assessments Richard K. Slocum * and Christopher E. Parrish School of Civil and Construction Engineering, Oregon State University, 101 Kearney Hall, 1491 SW Campus Way, Corvallis, OR 97331, USA; [email protected] * Correspondence: [email protected]; Tel.: +1-703-973-1983 Academic Editors: Gonzalo Pajares Martinsanz, Xiaofeng Li and Prasad S. Thenkabail Received: 13 March 2017; Accepted: 19 April 2017; Published: 22 April 2017 Abstract: Structure from motion (SfM) and MultiView Stereo (MVS) algorithms are increasingly being applied to imagery from unmanned aircraft systems (UAS) to generate point cloud data for various surveying and mapping applications. To date, the options for assessing the spatial accuracy of the SfM-MVS point clouds have primarily been limited to empirical accuracy assessments, which involve comparisons against reference data sets, which are both independent and of higher accuracy than the data they are being used to test. The acquisition of these reference data sets can be expensive, time consuming, and logistically challenging. Furthermore, these experiments are also almost always unable to be perfectly replicated and can contain numerous confounding variables, such as sun angle, cloud cover, wind, movement of objects in the scene, and camera thermal noise, to name a few. The combination of these factors leads to a situation in which robust, repeatable experiments are cost prohibitive, and the experiment results are frequently site-specific and condition-specific. Here, we present a workflow to render computer generated imagery using a virtual environment which can mimic the independent variables that would be experienced in a real-world UAS imagery acquisition scenario.
    [Show full text]
  • Objekttracking Für Dynamisches Videoprojektions Mapping
    Bachelorthesis Iwer Petersen Using object tracking for dynamic video projection mapping Fakultät Technik und Informatik Faculty of Engineering and Computer Science Studiendepartment Informatik Department of Computer Science Iwer Petersen Using object tracking for dynamic video projection mapping Bachelorthesis submitted within the scope of the Bachelor examination in the degree programme Bachelor of Science Technical Computer Science at the Department of Computer Science of the Faculty of Engineering and Computer Science of the Hamburg University of Applied Science Mentoring examiner: Prof. Dr. Ing. Birgit Wendholt Second expert: Prof. Dr.-Ing. Andreas Meisel Submitted at: January 31, 2013 Iwer Petersen Title of the paper Using object tracking for dynamic video projection mapping Keywords video projection mapping, object tracking, point cloud, 3D Abstract This document presents a way to realize video projection mapping onto moving objects. Therefore a visual 3D tracking method is used to determine the position and orientation of a known object. Via a calibrated projector-camera system, the real object then is augmented with a virtual texture. Iwer Petersen Thema der Arbeit Objekttracking für dynamisches Videoprojektions Mapping Stichworte Video Projektions Mapping, Objektverfolgung, Punktwolke, 3D Kurzzusammenfassung Dieses Dokument präsentiert einen Weg um Video Projektions Mapping auf sich bewegende Objekte zu realisieren. Dafür wird ein visuelles 3D Trackingverfahren verwendet um die Position und Lage eines bekannten Objekts zu bestimmen. Über ein kalibriertes Projektor- Kamera System wird das reale Objekt dann mit einer virtuellen Textur erweitert. Contents 1 Introduction1 1.1 Motivation . .1 1.2 Structure . .3 2 Related Work4 2.1 Corresponding projects . .4 2.1.1 Spatial Augmented Reality . .4 2.1.2 2D Mapping onto people .
    [Show full text]
  • Room Layout Estimation on Mobile Devices
    En vue de l'obtention du DOCTORAT DE L'UNIVERSITÉ DE TOULOUSE Délivré par : Institut National Polytechnique de Toulouse (Toulouse INP) Discipline ou spécialité : Image, Information et Hypermédia Présentée et soutenue par : M. VINCENT ANGLADON le vendredi 27 avril 2018 Titre : Room layout estimation on mobile devices Ecole doctorale : Mathématiques, Informatique, Télécommunications de Toulouse (MITT) Unité de recherche : Institut de Recherche en Informatique de Toulouse (I.R.I.T.) Directeur(s) de Thèse : M. VINCENT CHARVILLAT M. SIMONE GASPARINI Rapporteurs : M. CARSTEN GRIWODZ, UNIVERSITE D'OSLO M. DAVID FOFI, UNIVERSITE DE BOURGOGNE Membre(s) du jury : Mme LUCE MORIN, INSA DE RENNES, Président M. PASCAL BERTOLINO, UNIVERSITE GRENOBLE ALPES, Membre M. SIMONE GASPARINI, INP TOULOUSE, Membre M. TOMISLAV PRIBANIC, UNIVERSITE DE ZAGREB, Membre M. VINCENT CHARVILLAT, INP TOULOUSE, Membre Acknowledgments I would like to thank my thesis committee for giving me the opportunity to work on an exciting topic, and for the trust they placed in me. First, Simone Gasparini, I had the pleasure to have as advisor, who kept a cautious eye on my scientific and technical productions. I am certain this great attention to the details played an important role in the quality of my publications and the absence of rejection notice. Then, my thesis director, Vincent Charvillat, who was always generous in original ideas and positive energy. His advice helped me to put a more flattering light on my works and feel more optimistic. Finally, Telequid, which funded my works, with a special thought to Frédéric Bruel and Benjamin Ahsan for their great patience. I would also like to thank my referees: Prof.
    [Show full text]
  • GRASS GIS Loves Lidar FOSS4G NA 2016
    GRASS GIS loves lidar FOSS4G NA 2016 Vaclav Petras (Vashek) Anna Petrasova, Helena Mitasova Center for Geospatial Analytics May 5, 2016 available at wenzeslaus.github.io/grass-lidar-talks Vaclav Petras (NC State University) GRASS GIS loves lidar May 5, 2016 1 / 31 GRASS GIS I all in one I hydrology modeling, image segmentation, point clustering, . I from small laptops to supercomputers I Raspberry Pi, Windows, Mac, GNU/Linux, GNU/Hurd, FreeBSD, IBM AIX I learn now, use forever I over 30 years of development and interface refinement I probably used more than you think I similarly to C/C++ is often not latest release 7.0.4 mentioned but is somewhere in there Sunday, May 1, 2016 Vaclav Petras (NC State University) GRASS GIS loves lidar May 5, 2016 2 / 31 CLI Vaclav Petras (NC State University) GRASS GIS loves lidar May 5, 2016 3 / 31 GUI Vaclav Petras (NC State University) GRASS GIS loves lidar May 5, 2016 4 / 31 Python Vaclav Petras (NC State University) GRASS GIS loves lidar May 5, 2016 5 / 31 Python versus CLI Documentation, Command Line (Shell, Bash, cmd.exe): r.in.lidar input=points.las\ output=elevation-e Python: from grass.script import run_command run_command('r.in.lidar', input="points.las", output="elevation", flags='e') Vaclav Petras (NC State University) GRASS GIS loves lidar May 5, 2016 6 / 31 Module GUI Vaclav Petras (NC State University) GRASS GIS loves lidar May 5, 2016 7 / 31 Graphical Modeler Vaclav Petras (NC State University) GRASS GIS loves lidar May 5, 2016 8 / 31 Points I collected by lidar I generated by Structure
    [Show full text]
  • Basic Implementation and Measurements of Plane Detection
    Basic Implementation and Measurements of Plane Detection in Point Clouds A project of the 2017 Robotics Course of the School of Information Science and Technology (SIST) of ShanghaiTech University https://robotics.shanghaitech.edu.cn/teaching/robotics2017 Chen Chen Wentao Lv Yuan Yuan School of Information and School of Information and School of Information and Science Technology Science Technology Science Technology [email protected] [email protected] [email protected] ShanghaiTech University ShanghaiTech University ShanghaiTech University Abstract—In practical robotics research, plane detection is an Corsini et al.(2012) is working on dealing with the problem important prerequisite to a wide variety of vision tasks. FARO is of taking random samples over the surface of a 3D mesh another powerful devices to scan the whole environment into describing and evaluating efficient algorithms for generating point cloud. In our project, we are working to apply some algorithms to convert point cloud data to plane mesh, and then different distributions. In this paper, the author we propose path planning based on the plane information. Constrained Poisson-disk sampling, a new Poisson-disk sam- pling scheme for polygonal meshes which can be easily I. Introduction tweaked in order to generate customized set of points such as In practical robotics research, plane detection is an im- importance sampling or distributions with generic geometric portant prerequisite to a wide variety of vision tasks. Plane constraints. detection means to detect the plane information from some Kazhdan and Hoppe(2013) is talking about poisson surface basic disjoint information, for example, the point cloud. In this reconstruction, which can create watertight surfaces from project, we will use point clouds from FARO devices which oriented point sets.
    [Show full text]
  • Point Cloud Subjective Evaluation Methodology Based on Reconstructed Surfaces
    Point cloud subjective evaluation methodology based on reconstructed surfaces Evangelos Alexioua, Antonio M. G. Pinheirob, Carlos Duartec, Dragan Matkovi´cd, Emil Dumi´cd, Luis A. da Silva Cruzc, Lovorka Gotal Dmitrovi´cd, Marco V. Bernardob Manuela Pereirab and Touradj Ebrahimia aMultimedia Signal Processing Group, Ecole´ Polytechnique F´ed´eralede Lausanne, Switzerland; bInstitute of Telecommunications, University of Beira Interior, Portugal; cDepartment of Electrical and Computer Engineering, University of Coimbra and Institute of Telecommunications, Portugal; dDepartment of Electrical Engineering, University North, Croatia ABSTRACT Point clouds have been gaining importance as a solution to the problem of efficient representation of 3D geometric and visual information. They are commonly represented by large amounts of data, and compression schemes are important for their manipulation transmission and storing. However, the selection of appropriate compression schemes requires effective quality evaluation. In this work a subjective quality evaluation of point clouds using a surface representation is analyzed. Using a set of point cloud data objects encoded with the popular octree pruning method with different qualities, a subjective evaluation was designed. The point cloud geometry was presented to observers in the form of a movie showing the 3D Poisson reconstructed surface without textural information with the point of view changing in time. Subjective evaluations were performed in three different laboratories. Scores obtained from each test were correlated and no statistical differences were observed. Scores were also correlated with previous subjective tests and a good correlation was obtained when compared with mesh rendering in 2D monitors. Moreover, the results were correlated with state of the art point cloud objective metrics revealing poor correlation.
    [Show full text]
  • MEPP2: a Generic Platform for Processing 3D Meshes and Point Clouds
    EUROGRAPHICS 2020/ F. Banterle and A. Wilkie Short Paper MEPP2: a generic platform for processing 3D meshes and point clouds Vincent Vidal, Eric Lombardi , Martial Tola , Florent Dupont and Guillaume Lavoué Université de Lyon, CNRS, LIRIS, Lyon, France Abstract In this paper, we present MEPP2, an open-source C++ software development kit (SDK) for processing and visualizing 3D surface meshes and point clouds. It provides both an application programming interface (API) for creating new processing filters and a graphical user interface (GUI) that facilitates the integration of new filters as plugins. Static and dynamic 3D meshes and point clouds with appearance-related attributes (color, texture information, normal) are supported. The strength of the platform is to be generic programming oriented. It offers an abstraction layer, based on C++ Concepts, that provides interoperability over several third party mesh and point cloud data structures, such as OpenMesh, CGAL, and PCL. Generic code can be run on all data structures implementing the required concepts, which allows for performance and memory footprint comparison. Our platform also permits to create complex processing pipelines gathering idiosyncratic functionalities of the different libraries. We provide examples of such applications. MEPP2 runs on Windows, Linux & Mac OS X and is intended for engineers, researchers, but also students thanks to simple use, facilitated by the proposed architecture and extensive documentation. CCS Concepts • Computing methodologies ! Mesh models; Point-based models; • Software and its engineering ! Software libraries and repositories; 1. Introduction GUI for processing and visualizing 3D surface meshes and point clouds. With the increasing capability of 3D data acquisition devices, mod- Several platforms exist for processing 3D meshes such as Meshlab eling software and graphics processing units, three-dimensional [CCC08] based on VCGlibz, MEPP [LTD12], and libigl [JP∗18].
    [Show full text]