Depth Perception 0

Total Page:16

File Type:pdf, Size:1020Kb

Depth Perception 0 Depth Perception 0 Depth Perception Suggested reading:! ! Stephen E. Palmer (1999) Vision Science. ! Read JCA (2005) Early computational processing in binocular vision and depth perception. Progress in Biophysics and Molecular Biology 87: 77-108 Depth Perception 1 Depth Perception! Contents: Contents: • The problem of depth perception • • WasDepth is lerning cues ? • • ZellulärerPanum’s Mechanismus fusional area der LTP • • UnüberwachtesBinocular disparity Lernen • • ÜberwachtesRandom dot Lernen stereograms • • Delta-LernregelThe correspondence problem • • FehlerfunktionEarly models • • GradientenabstiegsverfahrenDisparity sensitive cells in V1 • • ReinforcementDisparity tuning Lernen curve • Binocular engergy model • The role of higher visual areas Depth Perception 2 The problem of depth perception! Our 3 dimensional environment gets reduced to two-dimensional images on the two retinae. Viewing Retinal geometry image The “lost” dimension is commonly called depth. As depth is crucial for almost all of our interaction with our environment and we perceive the 2D images as three dimensional the brain must try to estimate the depth from the retinal images as good as it can with the help of different cues and mechanisms. Depth Perception 3 Depth cues - monocular! Texture Accretion/Deletion Accommodation (monocular, optical) (monocular, ocular) Motion Parallax Convergence of Parallels (monocular, optical) (monocular, optical) Depth Perception 4 Depth cues - monocular ! Position relative to Horizon Familiar Size (monocular, optical) (monocular, optical) Relative Size Texture Gradients (monocular, optical) (monocular, optical) Depth Perception 5 Depth cues - monocular ! ! Edge Interpretation (monocular, optical) ! Shading and Shadows (monocular, optical) ! Aerial Perspective (monocular, optical) Depth Perception 6 Depth cues – binocular! Convergence Binocular Disparity (binocular, ocular) (binocular, optical) As the eyes are horizontally displaced we see the world from two slightly different viewpoints. The visual fields of the two eyes overlap in the central region of vision but points that are not on the so-called horopter fall onto different retinal positions. This lateral displacement which is relative to the fixation point is called binocular disparity. Depth Perception 7 Panum's fusional area! If we see different views of the world with both eyes why don't we experience double vision? Points in an area near the horopter (called Panum's fusional area) are fused perceptually into a single experienced image. Depth Perception 8 Panum's fusional area! You can easily check this perceptual fact for yourself Hold one forefinger in front of you and the other forefinger 30cm behind it. If you focus on one finger you will see the other finger twice because it is not within the limits of Panum's fusional region. Depth Perception 9 Binocular disparity! The point that we fixate on (here P) by definition has no binocular disparity as it falls directly onto the fovea at the same retinal position in both eyes. Fixating a point P the image of points like C that are closer to the viewer than P are displaced outwardly on the two retinae in a so-called crossed disparity, whereas farther points as F are displaced inwardly in a so-called uncrossed disparity. The direction and amount of binocular disparity therefore specify relative depth information in regard to the horopter. Depth Perception 10 Binocular disparity! Example of the use of binocular disparity for the experience of depth - Anaglyphs Nothing new, discovered in the 1850s by Joseph D’Almeida and Louis Du Hauron, William Friese-Green created the first 3D anaglyphic motion pictures in 1889. Depth Perception 11 Random dot stereograms! Until about 40 years ago, a viable theory was that the visual system operates on the two retinal images separately, performing scene segmentation and object recognition, and then finally, for each object in the scene, compares where its images fall in the two retinae. However, in the 1960s, Bela Julesz drew attention to the fact that stereopsis works perfectly well with random dot stereograms. Depth Perception 12 Random dot stereograms! Julesz, 1971 1. two identical images with random dots 2. cut out a shape that you want to appear in a different depth plane and slide it to the right in one image – the amount will determine the depth 3. fill the empty hole that has appeared through the sliding with random dots 4. put the original image left to the newly created one – that’s it! Random dot stereograms show that binocular disparity is such a strong depth cue that it does not need any monocular cue for depth perception. Depth Perception 13 The correspondence problem! Something is missing – the correspondence problem! Until now we have talked about how the direction and amount of disparity between corresponding image features in the two eyes can encode depth information but the problem we have excluded until now is how to determine which features in one retinal image correspond to which features in the other? Computer Vision: match contents of image parts using global cost functions. Depth Perception 14 The brains solution to correspondence problem! Depth Perception 15 Early models! First Marr-Poggio Algorithm (MIT 1977) Approach: matching individual pixels in the left and right image, inverse projection Constraints 1. Color constancy: corresponding points have the same color 2. Surface opacity: most surfaces in the world are opaque so only the nearest can be seen ! If A10 is a correct match, then B10, C10, D10 etc. cannot be correct. 3. Surface continuity: surfaces in the world tend to be locally continuous in depth ! excitatory connections between close matches Note! The constraints are heuristic assumptions that are Depth Perception 16 Early models! First Marr-Poggio Algorithm – dynamic behavior of the network Depth Perception 17 Disparity sensitive cells in V1! In the human visual system the inputs of the two eyes converge upon single cells in V1. Position disparity Phase disparity , Neuron, 1997. Qian From: Recent studies indicate that most disparity sensitive cells are hybrid, showing both position and phase disparity (Ohzawa, Freeman 1997/1999 – Livingstone, Tsao 1999 – Prince, Cumming, Parker 2002). Depth Perception 18 Disparity tuning curve! Schematic Drawing of a Disparity Tuning Curve for a Binocular Cell: Three representative retinal stimulus configurations correspond to a point in space that is nearer than, at, or further away from the fixation point. Depth Perception 19 Disparity tuning curve! Idealized tuning curves after characterization by Poggio & Fischer (1977) Disparity tuning: tuned-excitatory, near, tuned-inhibitory, and far cells. In each graph, the horizontal axis is disparity and the vertical axis is response rate. The curves are Gabor functions (the product of a cosine and a Gaussian); the angle represents the phase of the cosine component. Depth Perception 20 Binocular Energy model! Ohzawa et al., 1990; Qian, 1994, Read and Cumming 2007 A complex cell Cx receives input from binocular simple cells (BS). Receptive fields of two BS that differ in their phase by 90° get combined to a complex cell Cx. BS = Cx = BS + BS Depth Perception 21 Binocular Energy model ! Image value at position (x,y) Receptive field of equivalent left eye simple cell With the Gabor filter preferred preferred position phase disparity disparity with standard and deviation Depth Perception 22 Solution to the correspondence problem! V1 neurons are sensitive to local matches, not to global solutions of the correspondence problem. Anti-correlated stereograms With the discovery of V1 disparity sensitive cells the question arose if the correspondence problem was solved in V1 already or in a higher brain region. One approach to test for this were anti-correlated stereograms. These consist of normal stereograms but the brightness information in one of the images is reversed in comparison to the other image. Because of this there is no global- match solution to the stereo correspondence problem in anti-correlated stereograms. Nevertheless the disparity sensitive cells in V1 respond to the local false matches in anti-correlated stereograms which is a strong indication that the correspondence problem does not get solved in V1. Depth Perception 23 The role of higher visual areas! V2: relative disparity and disparity-defined edges (Thomas et al. 2002, von der Heydt et al., 2000, Qiu & von der Heydt, 2005) V5/MT: Microstimulation can bias the monkey’s judgement in a far-near stereo task that relies on absolute disparity (DeAngelis et al. 1998), but no strong support for other aspects of stereovision. V4: sensitive to the orientation of disparity defined planes (Hinkle & Connor, 2005). Depth Perception 24 Vergence Control! Depth Perception 25 Additional reading:! • Qian N. (1997) Binocular disparity and the perception of depth. Neuron, 18:359-68.! • Qian N, Zhu Y. (1997) Physiological computation of binocular disparity. Vision Res, 37:1811-27. ! Depth Perception 26 References:! • Hinkle, D.A. and Connor, C.E. (2005) Quantitative characterization of disparity tuning in ventral pathway area V4. Journal of Neurophysiology, 94:2726-37. • Ohzawa I, DeAngelis GC, and Freeman RD (1990) Stereoscopic depth discrimination in the visual cortex: neurons ideally suited as disparity detectors. Science 249:1037-1041. • Poggio GF, Fischer B. (1977) Binocular interaction and depth sensitivity in striate and prestriate cortex of behaving rhesus monkey. J Neurophysiol, 40:1392-405. • Qian, N (1994) Computing stereo disparity and motion with known binocular cell properties, Neural Computation, 6: 390-404.! • Qiu, F.T. and von der Heydt, R. (2005) Figure and ground in the visual cortex: V2 combines stereoscopic cues with Gestalt rules. Neuron, 47:155-166.! • Read JCA, Cumming BG (2007) Sensors for impossible stimuli may solve the stereo correspondence problem. Nature Neuroscience, 10:1322-1328. ! • Thomas OM, Cumming BG, and Parker AJ (2002) A specialization for relative disparity in V2. Nat Neurosci 5: 472–478.! • von der Heydt, R., Zhou, H. and Friedman, H.S. (2000) Representation of stereoscopic edges in monkey visual cortex. Vision Research, 40:1955-1967.!.
Recommended publications
  • Binocular Disparity - Difference Between Two Retinal Images
    Depth II + Motion I Lecture 13 (Chapters 6+8) Jonathan Pillow Sensation & Perception (PSY 345 / NEU 325) Spring 2015 1 depth from focus: tilt shift photography Keith Loutit : tilt shift + time-lapse photography http://vimeo.com/keithloutit/videos 2 Monocular depth cues: Pictorial Non-Pictorial • occlusion • motion parallax relative size • accommodation shadow • • (“depth from focus”) • texture gradient • height in plane • linear perspective 3 • Binocular depth cue: A depth cue that relies on information from both eyes 4 Two Retinas Capture Different images 5 Finger-Sausage Illusion: 6 Pen Test: Hold a pen out at half arm’s length With the other hand, see how rapidly you can place the cap on the pen. First using two eyes, then with one eye closed 7 Binocular depth cues: 1. Vergence angle - angle between the eyes convergence divergence If you know the angles, you can deduce the distance 8 Binocular depth cues: 2. Binocular Disparity - difference between two retinal images Stereopsis - depth perception that results from binocular disparity information (This is what they’re offering in “3D movies”...) 9 10 Retinal images in left & right eyes Figuring out the depth from these two images is a challenging computational problem. (Can you reason it out?) 11 12 Horopter: circle of points that fall at zero disparity (i.e., they land on corresponding parts of the two retinas) A bit of geometric reasoning will convince you that this surface is a circle containing the fixation point and the two eyes 13 14 point with uncrossed disparity point with crossed
    [Show full text]
  • Relationship Between Binocular Disparity and Motion Parallax in Surface Detection
    Perception & Psychophysics 1997,59 (3),370-380 Relationship between binocular disparity and motion parallax in surface detection JESSICATURNERand MYRON L, BRAUNSTEIN University ofCalifornia, Irvine, California and GEORGEJ. ANDERSEN University ofColifornia, Riverside, California The ability to detect surfaces was studied in a multiple-cue condition in which binocular disparity and motion parallax could specify independent depth configurations, On trials on which binocular disparity and motion parallax were presented together, either binocular disparity or motion parallax could indicate a surface in one of two intervals; in the other interval, both sources indicated a vol­ ume of random points. Surface detection when the two sources of information were present and compatible was not betterthan detection in baseline conditions, in which only one source of informa­ tion was present. When binocular disparity and motion specified incompatible depths, observers' ability to detect a surface was severely impaired ifmotion indicated a surface but binocular disparity did not. Performance was not as severely degraded when binocular disparity indicated a surface and motion did not. This dominance of binocular disparity persisted in the presence of foreknowledge about which source of information would be relevant. The ability to detect three-dimensional (3-D) surfaces action ofdepth cues in determining the perceived shape has been studied previously using static stereo displays of objects has been a subject of considerable attention. (e.g., Uttal, 1985), motion parallax displays (Andersen, Biilthoffand Mallot (1988) discussed four ways in which 1996; Andersen & Wuestefeld, 1993), and structure­ depth information from different cues may be combined: from-motion (SFM) displays (Turner, Braunstein, & An­ accumulation, veto, disambiguation, and cooperation.
    [Show full text]
  • 3D Computational Modeling and Perceptual Analysis of Kinetic Depth Effects
    Computational Visual Media DOI 10.1007/s41095-xxx-xxxx-x Vol. x, No. x, month year, xx–xx Research Article 3D Computational Modeling and Perceptual Analysis of Kinetic Depth Effects Meng-Yao Cui1, Shao-Ping Lu1( ), Miao Wang2, Yong-Liang Yang3, Yu-Kun Lai4, and Paul L. Rosin4 c The Author(s) 2020. This article is published with open access at Springerlink.com Abstract Humans have the ability to perceive 3D shapes from 2D projections of rotating 3D objects, which is called Kinetic Depth Effects. This process is based on a variety of visual cues such as lighting and shading effects. However, when such cues are weakened or missing, perception can become faulty, as demonstrated by the famous silhouette illusion example – the Spinning Dancer. Inspired by this, we establish objective and subjective evaluation models of rotated 3D objects by taking their projected 2D images as input. We investigate five different cues: ambient Fig. 1 The Spinning Dancer. Due to the lack of visual cues, it confuses humans luminance, shading, rotation speed, perspective, and color as to whether rotation is clockwise or counterclockwise. Here we show 3 of 34 difference between the objects and background. In the frames from the original animation [21]. Image courtesy of Nobuyuki Kayahara. objective evaluation model, we first apply 3D reconstruction algorithms to obtain an objective reconstruction quality 1 Introduction metric, and then use a quadratic stepwise regression analysis method to determine the weights among depth cues to The human perception mechanism of the 3D world has represent the reconstruction quality. In the subjective long been studied.
    [Show full text]
  • Interacting with Autostereograms
    See discussions, stats, and author profiles for this publication at: https://www.researchgate.net/publication/336204498 Interacting with Autostereograms Conference Paper · October 2019 DOI: 10.1145/3338286.3340141 CITATIONS READS 0 39 5 authors, including: William Delamare Pourang Irani Kochi University of Technology University of Manitoba 14 PUBLICATIONS 55 CITATIONS 184 PUBLICATIONS 2,641 CITATIONS SEE PROFILE SEE PROFILE Xiangshi Ren Kochi University of Technology 182 PUBLICATIONS 1,280 CITATIONS SEE PROFILE Some of the authors of this publication are also working on these related projects: Color Perception in Augmented Reality HMDs View project Collaboration Meets Interactive Spaces: A Springer Book View project All content following this page was uploaded by William Delamare on 21 October 2019. The user has requested enhancement of the downloaded file. Interacting with Autostereograms William Delamare∗ Junhyeok Kim Kochi University of Technology University of Waterloo Kochi, Japan Ontario, Canada University of Manitoba University of Manitoba Winnipeg, Canada Winnipeg, Canada [email protected] [email protected] Daichi Harada Pourang Irani Xiangshi Ren Kochi University of Technology University of Manitoba Kochi University of Technology Kochi, Japan Winnipeg, Canada Kochi, Japan [email protected] [email protected] [email protected] Figure 1: Illustrative examples using autostereograms. a) Password input. b) Wearable e-mail notification. c) Private space in collaborative conditions. d) 3D video game. e) Bar gamified special menu. Black elements represent the hidden 3D scene content. ABSTRACT practice. This learning effect transfers across display devices Autostereograms are 2D images that can reveal 3D content (smartphone to desktop screen). when viewed with a specific eye convergence, without using CCS CONCEPTS extra-apparatus.
    [Show full text]
  • Recovery of 3-D Shape from Binocular Disparity and Structure from Motion
    Perception & Psychophysics /993. 54 (2). /57-J(B Recovery of 3-D shape from binocular disparity and structure from motion JAMES S. TI'ITLE The Ohio State University, Columbus, Ohio and MYRON L. BRAUNSTEIN University of California, Irvine, California Four experiments were conducted to examine the integration of depth information from binocular stereopsis and structure from motion (SFM), using stereograms simulating transparent cylindri­ cal objects. We found that the judged depth increased when either rotational or translational motion was added to a display, but the increase was greater for rotating (SFM) displays. Judged depth decreased as texture element density increased for static and translating stereo displays, but it stayed relatively constant for rotating displays. This result indicates that SFM may facili­ tate stereo processing by helping to resolve the stereo correspondence problem. Overall, the re­ sults from these experiments provide evidence for a cooperative relationship betweel\..SFM and binocular disparity in the recovery of 3-D relationships from 2-D images. These findings indicate that the processing of depth information from SFM and binocular disparity is not strictly modu­ lar, and thus theories of combining visual information that assume strong modularity or indepen­ dence cannot accurately characterize all instances of depth perception from multiple sources. Human observers can perceive the 3-D shape and orien­ tion, both sources of information (along with many others) tation in depth of objects in a natural environment with im­ occur together in the natural environment. Thus, it is im­ pressive accuracy. Prior work demonstrates that informa­ portant to understand what interaction (if any) exists in the tion about shape and orientation can be recovered from visual processing of binocular disparity and SFM.
    [Show full text]
  • Stereopsis and Stereoblindness
    Exp. Brain Res. 10, 380-388 (1970) Stereopsis and Stereoblindness WHITMAN RICHARDS Department of Psychology, Massachusetts Institute of Technology, Cambridge (USA) Received December 20, 1969 Summary. Psychophysical tests reveal three classes of wide-field disparity detectors in man, responding respectively to crossed (near), uncrossed (far), and zero disparities. The probability of lacking one of these classes of detectors is about 30% which means that 2.7% of the population possess no wide-field stereopsis in one hemisphere. This small percentage corresponds to the probability of squint among adults, suggesting that fusional mechanisms might be disrupted when stereopsis is absent in one hemisphere. Key Words: Stereopsis -- Squint -- Depth perception -- Visual perception Stereopsis was discovered in 1838, when Wheatstone invented the stereoscope. By presenting separate images to each eye, a three dimensional impression could be obtained from two pictures that differed only in the relative horizontal disparities of the components of the picture. Because the impression of depth from the two combined disparate images is unique and clearly not present when viewing each picture alone, the disparity cue to distance is presumably processed and interpreted by the visual system (Ogle, 1962). This conjecture by the psychophysicists has received recent support from microelectrode recordings, which show that a sizeable portion of the binocular units in the visual cortex of the cat are sensitive to horizontal disparities (Barlow et al., 1967; Pettigrew et al., 1968a). Thus, as expected, there in fact appears to be a physiological basis for stereopsis that involves the analysis of spatially disparate retinal signals from each eye. Without such an analysing mechanism, man would be unable to detect horizontal binocular disparities and would be "stereoblind".
    [Show full text]
  • Relative Importance of Binocular Disparity and Motion Parallax for Depth Estimation: a Computer Vision Approach
    remote sensing Article Relative Importance of Binocular Disparity and Motion Parallax for Depth Estimation: A Computer Vision Approach Mostafa Mansour 1,2 , Pavel Davidson 1,* , Oleg Stepanov 2 and Robert Piché 1 1 Faculty of Information Technology and Communication Sciences, Tampere University, 33720 Tampere, Finland 2 Department of Information and Navigation Systems, ITMO University, 197101 St. Petersburg, Russia * Correspondence: pavel.davidson@tuni.fi Received: 4 July 2019; Accepted: 20 August 2019; Published: 23 August 2019 Abstract: Binocular disparity and motion parallax are the most important cues for depth estimation in human and computer vision. Here, we present an experimental study to evaluate the accuracy of these two cues in depth estimation to stationary objects in a static environment. Depth estimation via binocular disparity is most commonly implemented using stereo vision, which uses images from two or more cameras to triangulate and estimate distances. We use a commercial stereo camera mounted on a wheeled robot to create a depth map of the environment. The sequence of images obtained by one of these two cameras as well as the camera motion parameters serve as the input to our motion parallax-based depth estimation algorithm. The measured camera motion parameters include translational and angular velocities. Reference distance to the tracked features is provided by a LiDAR. Overall, our results show that at short distances stereo vision is more accurate, but at large distances the combination of parallax and camera motion provide better depth estimation. Therefore, by combining the two cues, one obtains depth estimation with greater range than is possible using either cue individually.
    [Show full text]
  • Chromostereo.Pdf
    ChromoStereoscopic Rendering for Trichromatic Displays Le¨ıla Schemali1;2 Elmar Eisemann3 1Telecom ParisTech CNRS LTCI 2XtremViz 3Delft University of Technology Figure 1: ChromaDepth R glasses act like a prism that disperses incoming light and induces a differing depth perception for different light wavelengths. As most displays are limited to mixing three primaries (RGB), the depth effect can be significantly reduced, when using the usual mapping of depth to hue. Our red to white to blue mapping and shading cues achieve a significant improvement. Abstract The chromostereopsis phenomenom leads to a differing depth per- ception of different color hues, e.g., red is perceived slightly in front of blue. In chromostereoscopic rendering 2D images are produced that encode depth in color. While the natural chromostereopsis of our human visual system is rather low, it can be enhanced via ChromaDepth R glasses, which induce chromatic aberrations in one Figure 2: Chromostereopsis can be due to: (a) longitunal chro- eye by refracting light of different wavelengths differently, hereby matic aberration, focus of blue shifts forward with respect to red, offsetting the projected position slightly in one eye. Although, it or (b) transverse chromatic aberration, blue shifts further toward might seem natural to map depth linearly to hue, which was also the the nasal part of the retina than red. (c) Shift in position leads to a basis of previous solutions, we demonstrate that such a mapping re- depth impression. duces the stereoscopic effect when using standard trichromatic dis- plays or printing systems. We propose an algorithm, which enables an improved stereoscopic experience with reduced artifacts.
    [Show full text]
  • Two Eyes See More Than One Human Beings Have Two Eyes Located About 6 Cm (About 2.4 In.) Apart
    ivi act ty 2 TTwowo EyesEyes SeeSee MoreMore ThanThan OneOne OBJECTIVES 1 straw 1 card, index (cut in half widthwise) Students discover how having two eyes helps us see in three dimensions and 3 pennies* increases our field of vision. 1 ruler, metric* The students For the class discover that each eye sees objects from a 1 roll string slightly different viewpoint to give us depth 1 roll tape, masking perception 1 pair scissors* observe that depth perception decreases *provided by the teacher with the use of just one eye measure their field of vision observe that their field of vision decreases PREPARATION with the use of just one eye Session I 1 Make a copy of Activity Sheet 2, Part A, for each student. SCHEDULE 2 Each team of two will need a metric ruler, Session I About 30 minutes three paper cups, and three pennies. Session II About 40 minutes Students will either close or cover their eyes, or you may use blindfolds. (Students should use their own blindfold—a bandanna or long strip VOCABULARY of cloth brought from home and stored in their science journals for use—in this depth perception and other activities.) field of vision peripheral vision Session II 1 Make a copy of Activity Sheet 2, Part B, for each student. MATERIALS 2 For each team, cut a length of string 50 cm (about 20 in.) long. Cut enough index For each student cards in half (widthwise) to give each team 1 Activity Sheet 2, Parts A and B half a card. Snip the corners of the cards to eliminate sharp edges.
    [Show full text]
  • Binocular Vision
    BINOCULAR VISION Rahul Bhola, MD Pediatric Ophthalmology Fellow The University of Iowa Department of Ophthalmology & Visual Sciences posted Jan. 18, 2006, updated Jan. 23, 2006 Binocular vision is one of the hallmarks of the human race that has bestowed on it the supremacy in the hierarchy of the animal kingdom. It is an asset with normal alignment of the two eyes, but becomes a liability when the alignment is lost. Binocular Single Vision may be defined as the state of simultaneous vision, which is achieved by the coordinated use of both eyes, so that separate and slightly dissimilar images arising in each eye are appreciated as a single image by the process of fusion. Thus binocular vision implies fusion, the blending of sight from the two eyes to form a single percept. Binocular Single Vision can be: 1. Normal – Binocular Single vision can be classified as normal when it is bifoveal and there is no manifest deviation. 2. Anomalous - Binocular Single vision is anomalous when the images of the fixated object are projected from the fovea of one eye and an extrafoveal area of the other eye i.e. when the visual direction of the retinal elements has changed. A small manifest strabismus is therefore always present in anomalous Binocular Single vision. Normal Binocular Single vision requires: 1. Clear Visual Axis leading to a reasonably clear vision in both eyes 2. The ability of the retino-cortical elements to function in association with each other to promote the fusion of two slightly dissimilar images i.e. Sensory fusion. 3. The precise co-ordination of the two eyes for all direction of gazes, so that corresponding retino-cortical element are placed in a position to deal with two images i.e.
    [Show full text]
  • Abstract Computer Vision in the Space of Light Rays
    ABSTRACT Title of dissertation: COMPUTER VISION IN THE SPACE OF LIGHT RAYS: PLENOPTIC VIDEOGEOMETRY AND POLYDIOPTRIC CAMERA DESIGN Jan Neumann, Doctor of Philosophy, 2004 Dissertation directed by: Professor Yiannis Aloimonos Department of Computer Science Most of the cameras used in computer vision, computer graphics, and image process- ing applications are designed to capture images that are similar to the images we see with our eyes. This enables an easy interpretation of the visual information by a human ob- server. Nowadays though, more and more processing of visual information is done by computers. Thus, it is worth questioning if these human inspired “eyes” are the optimal choice for processing visual information using a machine. In this thesis I will describe how one can study problems in computer vision with- out reference to a specific camera model by studying the geometry and statistics of the space of light rays that surrounds us. The study of the geometry will allow us to deter- mine all the possible constraints that exist in the visual input and could be utilized if we had a perfect sensor. Since no perfect sensor exists we use signal processing techniques to examine how well the constraints between different sets of light rays can be exploited given a specific camera model. A camera is modeled as a spatio-temporal filter in the space of light rays which lets us express the image formation process in a function ap- proximation framework. This framework then allows us to relate the geometry of the imaging camera to the performance of the vision system with regard to the given task.
    [Show full text]
  • Motion-In-Depth Perception and Prey Capture in the Praying Mantis Sphodromantis Lineola
    © 2019. Published by The Company of Biologists Ltd | Journal of Experimental Biology (2019) 222, jeb198614. doi:10.1242/jeb.198614 RESEARCH ARTICLE Motion-in-depth perception and prey capture in the praying mantis Sphodromantis lineola Vivek Nityananda1,*, Coline Joubier1,2, Jerry Tan1, Ghaith Tarawneh1 and Jenny C. A. Read1 ABSTRACT prey as they come near. Just as with depth perception, several cues Perceiving motion-in-depth is essential to detecting approaching could contribute to the perception of motion-in-depth. or receding objects, predators and prey. This can be achieved Two of the motion-in-depth cues that have received the most using several cues, including binocular stereoscopic cues such attention in humans are binocular: changing disparity and as changing disparity and interocular velocity differences, and interocular velocity differences (IOVDs) (Cormack et al., 2017). monocular cues such as looming. Although these have been Stereoscopic disparity refers to the difference in the position of an studied in detail in humans, only looming responses have been well object as seen by the two eyes. This disparity reflects the distance to characterized in insects and we know nothing about the role of an object. Thus as an object approaches, the disparity between the stereoscopic cues and how they might interact with looming cues. We two views changes. This changing disparity cue suffices to create a used our 3D insect cinema in a series of experiments to investigate perception of motion-in-depth for human observers, even in the the role of the stereoscopic cues mentioned above, as well as absence of other cues (Cumming and Parker, 1994).
    [Show full text]