Rapid 3D Measurement Using Digital Video Cameras

Total Page:16

File Type:pdf, Size:1020Kb

Rapid 3D Measurement Using Digital Video Cameras Rapid 3D Measurement Using Digital Video Cameras by Willem Johannes Van der Merwe March 2008 Rapid 3D Measurement Using Digital Video Cameras by Willem Johannes Van der Merwe Thesis presented in partial fulfillment of the requirements for the degree of Master of Science in Mechatronic Engineering at the University of Stellenbosch Supervisor: Dr. K. Schreve March 2008 Declaration I, the undersigned, hereby declare that the work contained in this thesis is my own original work and that I have not previously in its entirety or in part submitted it at any university for a degree. Signature: . W. J. Van der Merwe Date: . Copyright © 2008 University of Stellenbosch All rights reserved. i Abstract A rapid measurement system is implemented using two digital video cameras, presenting a faster and less expensive solution to certain metrology problems. The cameras are calibrated from one stereo image-pair of a 3D calibration grid that allows an immediate assessment of the achievable metric accuracy of the system. Three different methods, using either laser tracking or structured light patterns, were developed and employed to solve the coordinate extraction and correspondence matching problems. Different image processing techniques were used to speed up the entire measurement process. All software development was accomplished using only freely distributed software packages. The system achieves calibration in less than a minute and accumulates point correspondences at 12 frames per second. Accuracies of greater than 0.4 mm are achieved for a 235 x 190 x 95 mm measurement volume using a single pair of images with 640 x 480 pixel resolution each. ii Uittreksel Met die gebruik van twee digitale videokameras word ‘n spoedige meetsisteem ge їmplementeer om sodoende ‘n vinniger en meer bekostigbare oplossing te bied vir sekere metrologie probleme. Die kameras word gekalibreer deur een beeldpaar van ‘n 3D kalibrasieveld te gebruik wat oombliklike assessering van die behaalbare akkuraatheid van die sisteem moontlik maak. Drie metodes is ontwikkel en ge їmplementeer om die probleme van koördinaatontginning en puntooreenstemming op te los. Die laasgenoemde metodes maak gebruik van of ligpatrone of laser. Verskillende beeldverwerkingsmetodes word gebruik om die meetproses te verspoedig. Alle sagtewareontwikkeling is met vrylik beskikbare sagteware pakette gedoen. Die sisteem kan gekalibreer word in minder as ‘n minuut en versamel puntooreenstemmings teen ‘n tempo van 12 per sekonde. Akkuraatheid van beter as 0.4 mm word behaal vir ‘n meetvolume van 235 x 190 x 95 mm deur gebruik te maak van een paar beelde, elk met ‘n resolusie van 640 x 480. iii Acknowledgements My sincerest gratitude goes to the following people for helping to make this work possible: • Dr. Kristiaan Schreve, my supervisor, for his constant encouragement and competent guidance. • The members of the Machine Vision group, led by Prof. Ben Herbst and Dr. Karin Hunter, for freely sharing resources and ideas. Above all, I thank God, His Son and the Holy Spirit, through whom all things are made possible. iv Table of Contents Declaration...................................................................................................... i Abstract ......................................................................................................... ii Uittreksel.......................................................................................................iii Acknowledgements ..................................................................................... iv Table of Contents.......................................................................................... v List of Figures .............................................................................................. ix List of Tables................................................................................................ xi Abbreviations .............................................................................................. xii Chapter 1 Introduction.................................................................................. 1 1.1 Problem Statement..........................................................................................1 1.2 Project Context................................................................................................1 1.3 Thesis Outline .................................................................................................2 Chapter 2 Literature Review......................................................................... 3 2.1 Optical Measurement Techniques ...................................................................3 2.1.1 Passive Light Systems...........................................................................3 2.1.2 Active Light Systems .............................................................................4 2.2 Stereo Vision and Photogrammetry .................................................................5 2.3 Camera Calibration..........................................................................................6 2.3.1 Methods ................................................................................................6 2.3.2 Achieving Accuracy, Fast ......................................................................9 2.3.3 Common Factors Influencing Successful Calibration ...........................10 2.3.4 Calibration Object Design ....................................................................12 Chapter 3 The Camera Model..................................................................... 16 3.1 Pinhole Model................................................................................................16 3.1.1 Intrinsic Parameters.............................................................................17 v 3.1.2 Extrinsic Parameters ...........................................................................20 3.1.3 The Camera Matrix..............................................................................21 3.2 Additional Parameters ...................................................................................22 3.3 Lens Distortion Model....................................................................................22 Chapter 4 The Measurement Process ....................................................... 25 4.1 Camera Calibration........................................................................................25 4.1.1 The Method .........................................................................................25 4.1.2 Step 1: Initialisation of Camera Parameters.........................................25 4.1.3 Step 2: Refinement of the Camera Parameters ...................................26 4.2 Triangulation: Measuring Points in Three Dimensions ...................................30 4.2.1 The DLT Method..................................................................................30 4.2.2 Other Methods.....................................................................................30 4.3 Process Summary ......................................................................................31 Chapter 5 Image Processing...................................................................... 33 5.1 Software ........................................................................................................33 5.1.1 The Python Scripting Language...........................................................33 5.1.2 The OpenCV Software Package..........................................................34 5.1.3 Data Visualisation................................................................................35 5.1.4 C++ Code in Python: the Wrapping Principle.......................................36 5.1.5 Digital Video Camera Software............................................................36 5.2 Automation ....................................................................................................37 5.2.1 Pre-processing for Calibration .............................................................37 5.2.2 Point Correspondences in Two Images ...............................................38 5.3 Automated Detection of the Calibration Grid..................................................38 5.3.1 Assumptions........................................................................................38 5.3.2 Finding All Squares .............................................................................39 5.3.3 Intermediate Steps ..............................................................................42 5.3.4 Deriving Sub-pixel Coordinates for Square Corners ............................43 5.4 Rapid Correspondence Matching ..................................................................47 vi 5.4.1 Tracking a Moving Laser Dot...............................................................47 5.4.2 Corner Detection Using Square Projections.........................................49 5.4.3 Projected Line-Crossings.....................................................................51 Chapter 6 Hardware .................................................................................... 55 6.1 Digital Video Cameras...................................................................................55 6.1.1 Camera Properties and Characteristics ...............................................55 6.1.2 Lenses.................................................................................................56 6.1.3 Communication with the Computer......................................................57
Recommended publications
  • Depth Camera Technology Comparison and Performance Evaluation
    DEPTH CAMERA TECHNOLOGY COMPARISON AND PERFORMANCE EVALUATION Benjamin Langmann, Klaus Hartmann and Otmar Loffeld ZESS – Center for Sensor Systems, University of Siegen, Paul-Bonatz-Strasse 9-11, Siegen, Germany Keywords: Depth, Range camera, Time-of-Flight, ToF, Photonic Mixer Device, PMD, Comparison. Abstract: How good are cheap depth cameras, namely the Microsoft Kinect, compared to state of the art Time-of- Flight depth cameras? In this paper several depth cameras of different types were put to the test on a variety of tasks in order to judge their respective performance and to find out their weaknesses. We will concentrate on the common area of applications for which both types are specified, i.e. near field indoor scenes. The characteristics and limitations of the different technologies as well as the concrete hardware implementations are discussed and evaluated with a set of experimental setups. Especially, the noise level and the axial and angular resolutions are compared. Additionally, refined formulas to generate depth values based on the raw measurements of the Kinect are presented. 1 INTRODUCTION when used outdoors. Currently, ToF imaging chips reaching resolutions up to 200x200 pixels are on the Depth or range cameras have been developed for market and chips with 352x288 pixels are in several years and are available to researchers as well development and for a few years some ToF chips as on the market for certain applications for about a feature methods to suppress ambient light (e.g. decade. PMDTec, Mesa Imaging, 3DV Systems and Suppression of Background Illumination - SBI). Canesta were the companies driving the Other depth cameras or measuring devices, such development of Time-of-Flight (ToF) depth as laser scanners or structured light approaches, cameras.
    [Show full text]
  • Direct Interaction with Large Displays Through Monocular Computer Vision
    DIRECT INTERACTION WITH LARGE DISPLAYS THROUGH MONOCULAR COMPUTER VISION A thesis submitted in fulfilment of the requirements for the degree of Doctor of Philosophy in the School of Information Technologies at The University of Sydney Kelvin Cheng October 2008 © Copyright by Kelvin Cheng 2008 All Rights Reserved Abstract Large displays are everywhere, and have been shown to provide higher productivity gain and user satisfaction compared to traditional desktop monitors. The computer mouse remains the most common input tool for users to interact with these larger displays. Much effort has been made on making this interaction more natural and more intuitive for the user. The use of computer vision for this purpose has been well researched as it provides freedom and mobility to the user and allows them to interact at a distance. Interaction that relies on monocular computer vision, however, has not been well researched, particularly when used for depth information recovery. This thesis aims to investigate the feasibility of using monocular computer vision to allow bare-hand interaction with large display systems from a distance. By taking into account the location of the user and the interaction area available, a dynamic virtual touchscreen can be estimated between the display and the user. In the process, theories and techniques that make interaction with computer display as easy as pointing to real world objects is explored. Studies were conducted to investigate the way human point at objects naturally with their hand and to examine the inadequacy in existing pointing systems. Models that underpin the pointing strategy used in many of the previous interactive systems were formalized.
    [Show full text]
  • Scalable Multi-View Stereo Camera Array for Real World Real-Time Image Capture and Three-Dimensional Displays
    Scalable Multi-view Stereo Camera Array for Real World Real-Time Image Capture and Three-Dimensional Displays Samuel L. Hill B.S. Imaging and Photographic Technology Rochester Institute of Technology, 2000 M.S. Optical Sciences University of Arizona, 2002 Submitted to the Program in Media Arts and Sciences, School of Architecture and Planning in Partial Fulfillment of the Requirements for the Degree of Master of Science in Media Arts and Sciences at the Massachusetts Institute of Technology June 2004 © 2004 Massachusetts Institute of Technology. All Rights Reserved. Signature of Author:<_- Samuel L. Hill Program irlg edia Arts and Sciences May 2004 Certified by: / Dr. V. Michael Bove Jr. Principal Research Scientist Program in Media Arts and Sciences ZA Thesis Supervisor Accepted by: Andrew Lippman Chairperson Department Committee on Graduate Students MASSACHUSETTS INSTITUTE OF TECHNOLOGY Program in Media Arts and Sciences JUN 172 ROTCH LIBRARIES Scalable Multi-view Stereo Camera Array for Real World Real-Time Image Capture and Three-Dimensional Displays Samuel L. Hill Submitted to the Program in Media Arts and Sciences School of Architecture and Planning on May 7, 2004 in Partial Fulfillment of the Requirements for the Degree of Master of Science in Media Arts and Sciences Abstract The number of three-dimensional displays available is escalating and yet the capturing devices for multiple view content are focused on either single camera precision rigs that are limited to stationary objects or the use of synthetically created animations. In this work we will use the existence of inexpensive digital CMOS cameras to explore a multi- image capture paradigm and the gathering of real world real-time data of active and static scenes.
    [Show full text]
  • The Role of Camera Convergence in Stereoscopic Video See-Through Augmented Reality Displays
    (IJACSA) International Journal of Advanced Computer Science and Applications, Vol. 9, No. 8, 2018 The Role of Camera Convergence in Stereoscopic Video See-through Augmented Reality Displays Fabrizio Cutolo Vincenzo Ferrari University of Pisa University of Pisa Dept. of Information Engineering & EndoCAS Center Dept. of Information Engineering & EndoCAS Center Via Caruso 16, 56122, Pisa Via Caruso 16, 56122, Pisa Abstract—In the realm of wearable augmented reality (AR) merged with camera images captured by a stereo camera rig systems, stereoscopic video see-through displays raise issues rigidly fixed on the 3D display. related to the user’s perception of the three-dimensional space. This paper seeks to put forward few considerations regarding the The pixel-wise video mixing technology that underpins the perceptual artefacts common to standard stereoscopic video see- video see-through paradigm can offer high geometric through displays with fixed camera convergence. Among the coherence between virtual and real content. Nevertheless, the possible perceptual artefacts, the most significant one relates to industrial pioneers, as well as the early adopters of AR diplopia arising from reduced stereo overlaps and too large technology properly considered the camera-mediated view screen disparities. Two state-of-the-art solutions are reviewed. typical of video see-through devices as drastically affecting The first one suggests a dynamic change, via software, of the the quality of the visual perception and experience of the real virtual camera convergence,
    [Show full text]
  • Real-Time 3D Head Position Tracker System with Stereo
    REAL-TIME 3D HEAD POSITION TRACKER SYSTEM WITH STEREO CAMERAS USING A FACE RECOGNITION NEURAL NETWORK BY JAVIER IGNACIO GIRADO B. Electronics Engineering, ITBA University, Buenos Aires, Argentina, 1982 M. Electronics Engineering, ITBA University, Buenos Aires, Argentina 1984 THESIS Submitted as partial fulfillment of the requirements for the degree of Doctor of Philosophy in Computer Science in the Graduate College of the University of Illinois at Chicago, 2004 Chicago, Illinois ACKNOWLEDGMENTS I arrived at UIC in the winter of 1996, more than seven years ago. I had a lot to learn: how to find my way around a new school, a new city, a new country, a new culture and how to do computer science research. Those first years were very difficult for me and honestly I would not have made it if it were not for my old friends and the new ones I made at the laboratory and my colleagues. There are too many to list them all, so let me juts mention a few examples. I would like to thank my thesis committee (Thomas DeFanti, Daniel Sandin, Andrew Johnson, Jason Leigh and Joel Mambretti) for their unwavering support and assistance. They provided guidance in several areas that helped me accomplish my research goals and were very encouraging throughout the process. I would especially like to thank my thesis advisor Daniel Sandin for laying the foundation of this work and his continuous encouragement and feedback. He has been a constant source of great ideas and useful guidelines during my Thesis’ program. Thanks to Professor J. Ben-Arie for teaching me about science and computer vision.
    [Show full text]
  • 3D-Con2017program.Pdf
    There are 500 Stories at 3D-Con: This is One of Them I would like to welcome you to 3D-Con, a combined convention for the ISU and NSA. This is my second convention that I have been chairman for and fourth Southern California one that I have attended. Incidentally the first convention I chaired was the first one that used the moniker 3D-Con as suggested by Eric Kurland. This event has been harder to plan due to the absence of two friends who were movers and shakers from the last convention, David Washburn and Ray Zone. Both passed before their time soon after the last convention. I thought about both often when planning for this convention. The old police procedural movie the Naked City starts with the quote “There are eight million stories in the naked city; this has been one of them.” The same can be said of our interest in 3D. Everyone usually has an interesting and per- sonal reason that they migrated into this unusual hobby. In Figure 1 My Dad and his sister on a keystone view 1932. a talk I did at the last convention I mentioned how I got inter- ested in 3D. I was visiting the Getty Museum in southern Cali- fornia where they had a sequential viewer with 3D Civil War stereoviews, which I found fascinating. My wife then bought me some cards and a Holmes viewer for my birthday. When my family learned that I had a stereo viewer they sent me the only surviving photographs from my fa- ther’s childhood which happened to be stereoviews tak- en in 1932 in Norwalk, Ohio by the Keystone View Com- pany.
    [Show full text]
  • 3D Frequently Asked Questions
    3D Frequently Asked Questions Compiled from the 3-D mailing list 3D Frequently Asked Questions This document was compiled from postings on the 3D electronic mail group by: Joel Alpers For additions and corrections, please contact me at: [email protected] This is Revision 1.1, January 5, 1995 The information in this document is provided free of charge. You may freely distribute this document to anyone you desire provided that it is passed on unaltered with this notice intact, and that it be provided free of charge. You may charge a reasonable fee for duplication and postage. This information is deemed accurate but is not guaranteed. 2 Table Of Contents 1 Introduction . 7 1.1 The 3D mailing list . 7 1.2 3D Basics . 7 2 Useful References . 7 3 3D Time Line . 8 4 Suppliers . 9 5 Processing / Mounting . 9 6 3D film formats . 9 7 Viewing Stereo Pairs . 11 7.1 Free Viewing - viewing stereo pairs without special equipment . 11 7.1.1 Parallel viewing . 11 7.1.2 Cross-eyed viewing . 11 7.1.3 Sample 3D images . 11 7.2 Viewing using 3D viewers . 11 7.2.1 Print viewers - no longer manufactured, available used . 11 7.2.2 Print viewers - currently manufactured . 12 7.2.3 Slide viewers - no longer manufactured, available used . 12 7.2.4 Slide viewers - currently manufactured . 12 8 Stereo Cameras . 13 8.1 Currently Manufactured . 13 8.2 Available used . 13 8.3 Custom Cameras . 13 8.4 Other Techniques . 19 8.4.1 Twin Camera . 19 8.4.2 Slide Bar .
    [Show full text]
  • Changing Perspective in Stereoscopic Images
    1 Changing Perspective in Stereoscopic Images Song-Pei Du, Shi-Min Hu, Member, IEEE, and Ralph R. Martin Abstract—Traditional image editing techniques cannot be directly used to edit stereoscopic (‘3D’) media, as extra constraints are needed to ensure consistent changes are made to both left and right images. Here, we consider manipulating perspective in stereoscopic pairs. A straightforward approach based on depth recovery is unsatisfactory: instead, we use feature correspon- dences between stereoscopic image pairs. Given a new, user-specified perspective, we determine correspondence constraints under this perspective, and optimize a 2D warp for each image which preserves straight lines and guarantees proper stereopsis relative to the new camera. Experiments verify that our method generates new stereoscopic views which correspond well to expected projections, for a wide range of specified perspective. Various advanced camera effects, such as dolly zoom and wide angle effects, can also be readily generated for stereoscopic image pairs using our method. Index Terms—Stereoscopic images, stereopsis, perspective, warping. F 1 INTRODUCTION stereoscopic images, such as stereoscopic inpainting HE renaissance of ‘3D’ movies has led to the [6], disparity mapping [3], [7], and 3D copy & paste T development of both hardware and software [4]. However, relatively few tools are available com- stereoscopic techniques for both professionals and pared to the number of 2D image tools. consumers. Availability of 3D content has dramati- Perspective plays an essential role in the appearance cally increased, with 3D movies shown in cinemas, of images. In 2D, a simple approach to changing the launches of 3D TV channels, and 3D games on PCs.
    [Show full text]
  • Per-Pixel Calibration for RGB-Depth Natural 3D Reconstruction on GPU
    University of Kentucky UKnowledge Theses and Dissertations--Electrical and Computer Engineering Electrical and Computer Engineering 2016 Per-Pixel Calibration for RGB-Depth Natural 3D Reconstruction on GPU Sen Li University of Kentucky, [email protected] Author ORCID Identifier: http://orcid.org/0000-0001-9906-0481 Digital Object Identifier: https://doi.org/10.13023/ETD.2016.416 Right click to open a feedback form in a new tab to let us know how this document benefits ou.y Recommended Citation Li, Sen, "Per-Pixel Calibration for RGB-Depth Natural 3D Reconstruction on GPU" (2016). Theses and Dissertations--Electrical and Computer Engineering. 93. https://uknowledge.uky.edu/ece_etds/93 This Master's Thesis is brought to you for free and open access by the Electrical and Computer Engineering at UKnowledge. It has been accepted for inclusion in Theses and Dissertations--Electrical and Computer Engineering by an authorized administrator of UKnowledge. For more information, please contact [email protected]. STUDENT AGREEMENT: I represent that my thesis or dissertation and abstract are my original work. Proper attribution has been given to all outside sources. I understand that I am solely responsible for obtaining any needed copyright permissions. I have obtained needed written permission statement(s) from the owner(s) of each third-party copyrighted matter to be included in my work, allowing electronic distribution (if such use is not permitted by the fair use doctrine) which will be submitted to UKnowledge as Additional File. I hereby grant to The University of Kentucky and its agents the irrevocable, non-exclusive, and royalty-free license to archive and make accessible my work in whole or in part in all forms of media, now or hereafter known.
    [Show full text]
  • Motion Capture Technologies 24
    MOTION CAPTURE TECHNOLOGIES 24. 3. 2015 DISA Seminar Jakub Valcik Outline Human Motion and Digitalization Motion Capture Devices Optical Ranging Sensor Inertial and Magnetic Comparison Human Motion Brain + Skeleton + Muscles Internal factors External Factors Musculature Shoes Skeleton Clothes Weight Type of surface Injuries Slope of surface Movement habits Wind Pregnancy Gravity State of mind Environment Motion Capture 푛 Sequence of individual frames 푚 = (푓푘)푘=1 Captured information Static body configuration Position in space Orientation of body Silhouette vs. Skeleton Appearance Based Approaches Eadweard Muybridge, 1878 Boys Playing Leapfrog, printed 1887 Appearance Base Approaches Originally only regular video cameras, CCTV Silhouette oriented 푓푘 ≔ 퐵푘 푥, 푦 ∈ {0, 1} Silhouette extraction is a common bottleneck Model Based Approaches Additional abstraction – 2D, 3D Stick figure, volumetric model, … Skeleton Joint (or end-effector), Bone Undirected acyclic graph 퐽 – tree Joint ~ Vertex, Bone ~ Edge 3×|푉 퐽 | Static configuration ~ Pose 푝 ∈ ℝ 푛 푛 푚 = (푓푘)푘=1= (푝푘)푘=1 Motion Capture Devices Optical Both, appearance and Markerless model based approaches Invasive Inertial Magnetic Only model based Mechanical approaches Radio frequency Optical Markerless MoCap Devices 3D scene reconstruction Multiple views Depth sensors RGB stereoscopic cameras Multiple synchronized cameras Additional sensors Silhouette extraction Depth sensing IR camera, ranging sensor Ranging Sensor - Triangulation Laser
    [Show full text]
  • Distance Estimation from Stereo Vision: Review and Results
    DISTANCE ESTIMATION FROM STEREO VISION: REVIEW AND RESULTS A Project Presented to the faculty of the Department of the Computer Engineering California State University, Sacramento Submitted in partial satisfaction of the requirements for the degree of MASTER OF SCIENCE in Computer Engineering by Sarmad Khalooq Yaseen SPRING 2019 © 2019 Sarmad Khalooq Yaseen ALL RIGHTS RESERVED ii DISTANCE ESTIMATION FROM STEREO VISION: REVIEW AND RESULTS A Project by Sarmad Khalooq Yaseen Approved by: __________________________________, Committee Chair Dr. Fethi Belkhouche __________________________________, Second Reader Dr. Preetham B. Kumar ____________________________ Date iii Student: Sarmad Khalooq Yaseen I certify that this student has met the requirements for format contained in the University format manual, and that this project is suitable for shelving in the Library and credit is to be awarded for the project. __________________________, Graduate Coordinator ___________________ Dr. Behnam S. Arad Date Department of Computer Engineering iv Abstract of DISTANCE ESTIMATION FROM STEREO VISION: REVIEW AND RESULTS by Sarmad Khalooq Yaseen Stereo vision is the one of the major researched domains of computer vision, and it can be used for different applications, among them, extraction of the depth of a scene. This project provides a review of stereo vision and matching algorithms, used to solve the correspondence problem [22]. A stereo vision system consists of two cameras with parallel optical axes which are separated from each other by a small distance. The system is used to produce two stereoscopic pictures of a given object. Distance estimation between the object and the cameras depends on two factors, the distance between the positions of the object in the pictures and the focal lengths of the cameras [37].
    [Show full text]
  • Lecture 1: Course Introduction and Overview
    MIT Media Lab MAS 131/ 531 Computational Camera & Photography: Camera Culture Ramesh Raskar MIT Media Lab http://cameraculture.media.mit.edu/ MIT Media Lab Image removed due to copyright restrictions. Photo of airplane propeller, taken with iPhone and showing aliasing effect: http://scalarmotion.wordpress.com/2009/03/15/propeller-image-aliasing/ Shower Curtain: Diffuser Courtesy of Shree Nayar. Used with permission. Source: http://www1.cs.columbia.edu/CAVE/projects/ separation/occluders_gallery.php Direct Global A Teaser: Dual Photography Projector Photocell Figure by MIT OpenCourseWare. Images: Projector by MIT OpenCourseWare. Scene Photocell courtesy of afternoon_sunlight on Flickr. Scene courtesy of sbpoet on Flickr. A Teaser: Dual Photography Projector Photocell Figure by MIT OpenCourseWare. Images: Projector by MIT OpenCourseWare. Scene Photocell courtesy of afternoon_sunlight on Flickr. Scene courtesy of sbpoet on Flickr. A Teaser: Dual Photography Projector Camera Figure by MIT OpenCourseWare. Scene Images: Projector and camera by MIT OpenCourseWare. Scene courtesy of sbpoet on Flickr. The 4D transport matrix: Contribution of each projector pixel to each camera pixel projector camera Figure by MIT OpenCourseWare. Photo courtesy of sbpoet on Flickr. Scene Images: Projector and camera by MIT OpenCourseWare. Scene courtesy of sbpoet on Flickr. The 4D transport matrix: Contribution of each projector pixel to each camera pixel projector camera Figure by MIT OpenCourseWare. Scene Images: Projector and camera by MIT OpenCourseWare. Sen et al, Siggraph 2005 Scene courtesy of sbpoet on Flickr. The 4D transport matrix: Which projector pixel contributes to each camera pixel projector camera Figure by MIT OpenCourseWare. ? Scene Images: Projector and camera by MIT OpenCourseWare. Sen et al, Siggraph 2005 Scene courtesy of sbpoet on Flickr.
    [Show full text]