On the Importance of Stereo for Accurate Depth Estimation: an Efficient Semi-Supervised Deep Neural Network Approach

Total Page:16

File Type:pdf, Size:1020Kb

On the Importance of Stereo for Accurate Depth Estimation: an Efficient Semi-Supervised Deep Neural Network Approach On the Importance of Stereo for Accurate Depth Estimation: An Efficient Semi-Supervised Deep Neural Network Approach Nikolai Smolyanskiy Alexey Kamenev Stan Birchfield NVIDIA {nsmolyanskiy, akamenev, sbirchfield}@nvidia.com Abstract stereo cameras); and 2) multiple images provide geomet- ric constraints that can be leveraged to overcome the many We revisit the problem of visual depth estimation in the ambiguities of photometric data. context of autonomous vehicles. Despite the progress on The alternative is to use a single image to estimate monocular depth estimation in recent years, we show that depth. We argue that this alternative—due to its funda- the gap between monocular and stereo depth accuracy re- mental limitations—is not likely to be able to achieve high- mains large—a particularly relevant result due to the preva- accuracy depth estimation at large distances in unfamiliar lent reliance upon monocular cameras by vehicles that are environments. As a result, in the context of self-driving expected to be self-driving. We argue that the challenges cars we believe monocular depth estimation is not likely of removing this gap are significant, owing to fundamen- to yield results with sufficient accuracy. In contrast, we tal limitations of monocular vision. As a result, we focus offer a novel, efficient deep-learning stereo approach that our efforts on depth estimation by stereo. We propose a achieves compelling results on the KITTI 2015 dataset by novel semi-supervised learning approach to training a deep leveraging a semi-supervised loss function (using LIDAR stereo neural network, along with a novel architecture con- and photometric consistency), concatenating cost volume, taining a machine-learned argmax layer and a custom run- 3D convolutions, and a machine-learned argmax function. time that enables a smaller version of our stereo DNN to The contributions of the paper are as follows: run on an embedded GPU. Competitive results are shown on the KITTI 2015 stereo dataset. We also evaluate the re- • Quantitative and qualitative demonstration of the gap cent progress of stereo algorithms by measuring the impact in depth accuracy between monocular and stereoscopic upon accuracy of various design criteria.1 depth. • A novel semi-supervised approach (combining lidar and photometric losses) to training a deep stereo neu- 1. Introduction ral network. To our knowledge, ours is the first deep stereo network to do so.3 Estimating depth from images is a long-standing prob- lem in computer vision. Depth perception is useful • A smaller version of our network, and a custom run- for scene understanding, scene reconstruction, virtual and time, that runs at near real-time (∼20 fps) on a stan- augmented reality, obstacle avoidance, self-driving cars, dard GPU, and runs efficiently on an embedded GPU. robotics, and other applications. To our knowledge, ours is the first stereo DNN to run Traditionally, multiple images have been used to esti- on an embedded GPU. mate depth. Techniques that fall within this category in- • Quantitative analysis of various network design clude stereo, photometric stereo, depth from focus, depth choices, along with a novel machine-learned argmax from defocus, time-of-flight,2 and structure from motion. layer that yields smoother disparity maps. The reasons for using multiple images are twofold: 1) abso- lute depth estimates require at least one known distance in 2. Motivation the world, which can often be provided by some knowledge The undeniable success of deep neural networks in com- regarding the multi-camera rig (e.g., the baseline between puter vision has encouraged researchers to pursue the prob- 1Video of the system is at https://youtu.be/0FPQdVOYoAU. lem of estimating depth from a single image [5, 20, 6, 9, 17]. 2Although time-of-flight does not, in theory require multiple images, in practice multiple images are collected with different bandwidths in order 3Similarly, Kuznietsov et al. [17] use a semi-supervised approach for to achieve high accuracy over long ranges. training a monocular network. 1120 This is, no doubt, a noble endeavor: if it were possible to ac- engineering approach to such a safety-critical system is to curately estimate depth from a single image, then the com- leverage all available sensors rather than assume they are plexity (and hence cost) of the hardware needed would be not needed; thus, we believe that it is important to accu- dramatically reduced, which would broaden the applicabil- rately assess the increased error in depth estimation when ity substantially. An excellent overview of existing work on relying upon monocular cameras. monocular depth estimation can be found in [9]. At typical highway speeds, the braking distance re- Nevertheless, there are reasons to be cautious about the quired to completely stop before impact necessitates ob- reported success of monocular depth. To date, monocular serving an unforeseen stopped object approximately 100 m depth solutions, while yielding encouraging preliminary re- away. Intrigued by the reported success of monocular sults, are not at the point where reliable information (from a depth, we tried some recent algorithms, only to discover robotics point of view) can be expected from them. And al- that monocular depth is not able to achieve accuracies any- though such solutions will continue to improve, monocular where close to that requirement. We then turned our at- depth will never overcome well-known fundamental limi- tention to stereo, where significant progress has been made tations, such as the need for a world measurement to infer in recent years in applying deep learning to the problem absolute depth, and the ambiguity that arises when a pho- [25, 24, 11, 27, 26, 29, 8, 15, 23]. An excellent overview tograph is taken of a photograph (an important observation of recent stereo algorithms can be found in [15]. In this for biometric and security systems). flurry of activity, a variety of architectures have been pro- One of the motivations for monocular depth is a long- posed, but there has been no systematic study as to how standing belief that stereo is only useful at close range. these design choices impact quality. One purpose of this It has been widely reported, for example in [10], that be- paper is thus to investigate several of these options in order yond about 6 meters, the human visual system is essentially to quantify their impact, which we do in Sec. 5. In the con- monocular. But there is mounting evidence that the hu- text of this study, we develop a novel semi-supervised stereo man stereo system is actually much more capable than that. approach, which we present in Sec. 4. First, however, we il- Multiple studies have shown metric depth estimation up to lustrate the limitations of monocular depth estimation in the 20 meters [18, 1]; and, although error increases as disparity next section. increases [13], controlled experiments have confirmed that scaled disparity can be estimated up to 300 m, even with- 3. Difficulties of Monocular Depth Estimation out any depth cues from monocular vision [22]. Moreover, since the human visual system is capable of estimating dis- To appreciate the gap between mono and stereo vision, parity as small as a few seconds of arc [22], there is rea- consider the image of Fig. 1, with several points of interest son to believe that the distance could be 1 km or greater, highlighted. Without knowing the scene, if you were to ask with some evidence supporting such a claim provided by yourself whether the width of the near road (on which the the experiments of [4]. Note that an artificial stereo system car (A) sits) is greater than the width of the far tracks (dis- whose baseline is wider than the average 65 mm interpupil- tance between the near and far poles (E and F)), you might lary distance of the human visual system has the potential be tempted to answer in the affirmative. After all, the road to provide even greater accuracy. not only occupies more pixels in the image (which is to be This question takes on renewed significance in the con- expected, since it is closer to the camera), but it occupies text of self-driving cars, since most automobile manufac- orders of magnitude more pixels. We showed this image to turers and experimental autonomous vehicles do not install several people in our lab, and they all reached the same con- stereo cameras in their vehicles.4 Rather, these systems clusion: the road indeed appears to be significantly wider. rely on various combinations of monocular cameras, lidar, As it turns out, if this image is any indication, people are not 6 radar, and sonar sensors.5 For detecting static obstacles very good at estimating metric depth from a single image. such as trees, poles, railings, and concrete barriers, most The output of a leading monocular depth algorithm, systems rely on cameras and/or lidar. Although it is be- called MonoDepth [9], is shown in Fig. 2,7 along with the yond the scope of this paper whether monocular cameras are output of our stereo depth algorithm. At first glance, both sufficient for self-driving behavior (certainly people with results appear plausible. Although the stereo algorithm pre- monocular vision can drive safely in most situations), or serves crisper object boundaries and appears at least slightly whether stereo is better than lidar, we argue that the proper more accurate, it is difficult to tell from the grayscale im- 4To the authors’ knowledge, at the time of this writing stereo cameras 6Specifically, we asked 8 people to estimate the distance to the fence can be found only on certain models of Mercedes and Subaru vehicles; no (ground truth 14 m) and the distance to the building (ground truth 30 m).
Recommended publications
  • 3D Computational Modeling and Perceptual Analysis of Kinetic Depth Effects
    Computational Visual Media DOI 10.1007/s41095-xxx-xxxx-x Vol. x, No. x, month year, xx–xx Research Article 3D Computational Modeling and Perceptual Analysis of Kinetic Depth Effects Meng-Yao Cui1, Shao-Ping Lu1( ), Miao Wang2, Yong-Liang Yang3, Yu-Kun Lai4, and Paul L. Rosin4 c The Author(s) 2020. This article is published with open access at Springerlink.com Abstract Humans have the ability to perceive 3D shapes from 2D projections of rotating 3D objects, which is called Kinetic Depth Effects. This process is based on a variety of visual cues such as lighting and shading effects. However, when such cues are weakened or missing, perception can become faulty, as demonstrated by the famous silhouette illusion example – the Spinning Dancer. Inspired by this, we establish objective and subjective evaluation models of rotated 3D objects by taking their projected 2D images as input. We investigate five different cues: ambient Fig. 1 The Spinning Dancer. Due to the lack of visual cues, it confuses humans luminance, shading, rotation speed, perspective, and color as to whether rotation is clockwise or counterclockwise. Here we show 3 of 34 difference between the objects and background. In the frames from the original animation [21]. Image courtesy of Nobuyuki Kayahara. objective evaluation model, we first apply 3D reconstruction algorithms to obtain an objective reconstruction quality 1 Introduction metric, and then use a quadratic stepwise regression analysis method to determine the weights among depth cues to The human perception mechanism of the 3D world has represent the reconstruction quality. In the subjective long been studied.
    [Show full text]
  • Interacting with Autostereograms
    See discussions, stats, and author profiles for this publication at: https://www.researchgate.net/publication/336204498 Interacting with Autostereograms Conference Paper · October 2019 DOI: 10.1145/3338286.3340141 CITATIONS READS 0 39 5 authors, including: William Delamare Pourang Irani Kochi University of Technology University of Manitoba 14 PUBLICATIONS 55 CITATIONS 184 PUBLICATIONS 2,641 CITATIONS SEE PROFILE SEE PROFILE Xiangshi Ren Kochi University of Technology 182 PUBLICATIONS 1,280 CITATIONS SEE PROFILE Some of the authors of this publication are also working on these related projects: Color Perception in Augmented Reality HMDs View project Collaboration Meets Interactive Spaces: A Springer Book View project All content following this page was uploaded by William Delamare on 21 October 2019. The user has requested enhancement of the downloaded file. Interacting with Autostereograms William Delamare∗ Junhyeok Kim Kochi University of Technology University of Waterloo Kochi, Japan Ontario, Canada University of Manitoba University of Manitoba Winnipeg, Canada Winnipeg, Canada [email protected] [email protected] Daichi Harada Pourang Irani Xiangshi Ren Kochi University of Technology University of Manitoba Kochi University of Technology Kochi, Japan Winnipeg, Canada Kochi, Japan [email protected] [email protected] [email protected] Figure 1: Illustrative examples using autostereograms. a) Password input. b) Wearable e-mail notification. c) Private space in collaborative conditions. d) 3D video game. e) Bar gamified special menu. Black elements represent the hidden 3D scene content. ABSTRACT practice. This learning effect transfers across display devices Autostereograms are 2D images that can reveal 3D content (smartphone to desktop screen). when viewed with a specific eye convergence, without using CCS CONCEPTS extra-apparatus.
    [Show full text]
  • Stereopsis and Stereoblindness
    Exp. Brain Res. 10, 380-388 (1970) Stereopsis and Stereoblindness WHITMAN RICHARDS Department of Psychology, Massachusetts Institute of Technology, Cambridge (USA) Received December 20, 1969 Summary. Psychophysical tests reveal three classes of wide-field disparity detectors in man, responding respectively to crossed (near), uncrossed (far), and zero disparities. The probability of lacking one of these classes of detectors is about 30% which means that 2.7% of the population possess no wide-field stereopsis in one hemisphere. This small percentage corresponds to the probability of squint among adults, suggesting that fusional mechanisms might be disrupted when stereopsis is absent in one hemisphere. Key Words: Stereopsis -- Squint -- Depth perception -- Visual perception Stereopsis was discovered in 1838, when Wheatstone invented the stereoscope. By presenting separate images to each eye, a three dimensional impression could be obtained from two pictures that differed only in the relative horizontal disparities of the components of the picture. Because the impression of depth from the two combined disparate images is unique and clearly not present when viewing each picture alone, the disparity cue to distance is presumably processed and interpreted by the visual system (Ogle, 1962). This conjecture by the psychophysicists has received recent support from microelectrode recordings, which show that a sizeable portion of the binocular units in the visual cortex of the cat are sensitive to horizontal disparities (Barlow et al., 1967; Pettigrew et al., 1968a). Thus, as expected, there in fact appears to be a physiological basis for stereopsis that involves the analysis of spatially disparate retinal signals from each eye. Without such an analysing mechanism, man would be unable to detect horizontal binocular disparities and would be "stereoblind".
    [Show full text]
  • Relative Importance of Binocular Disparity and Motion Parallax for Depth Estimation: a Computer Vision Approach
    remote sensing Article Relative Importance of Binocular Disparity and Motion Parallax for Depth Estimation: A Computer Vision Approach Mostafa Mansour 1,2 , Pavel Davidson 1,* , Oleg Stepanov 2 and Robert Piché 1 1 Faculty of Information Technology and Communication Sciences, Tampere University, 33720 Tampere, Finland 2 Department of Information and Navigation Systems, ITMO University, 197101 St. Petersburg, Russia * Correspondence: pavel.davidson@tuni.fi Received: 4 July 2019; Accepted: 20 August 2019; Published: 23 August 2019 Abstract: Binocular disparity and motion parallax are the most important cues for depth estimation in human and computer vision. Here, we present an experimental study to evaluate the accuracy of these two cues in depth estimation to stationary objects in a static environment. Depth estimation via binocular disparity is most commonly implemented using stereo vision, which uses images from two or more cameras to triangulate and estimate distances. We use a commercial stereo camera mounted on a wheeled robot to create a depth map of the environment. The sequence of images obtained by one of these two cameras as well as the camera motion parameters serve as the input to our motion parallax-based depth estimation algorithm. The measured camera motion parameters include translational and angular velocities. Reference distance to the tracked features is provided by a LiDAR. Overall, our results show that at short distances stereo vision is more accurate, but at large distances the combination of parallax and camera motion provide better depth estimation. Therefore, by combining the two cues, one obtains depth estimation with greater range than is possible using either cue individually.
    [Show full text]
  • Chromostereo.Pdf
    ChromoStereoscopic Rendering for Trichromatic Displays Le¨ıla Schemali1;2 Elmar Eisemann3 1Telecom ParisTech CNRS LTCI 2XtremViz 3Delft University of Technology Figure 1: ChromaDepth R glasses act like a prism that disperses incoming light and induces a differing depth perception for different light wavelengths. As most displays are limited to mixing three primaries (RGB), the depth effect can be significantly reduced, when using the usual mapping of depth to hue. Our red to white to blue mapping and shading cues achieve a significant improvement. Abstract The chromostereopsis phenomenom leads to a differing depth per- ception of different color hues, e.g., red is perceived slightly in front of blue. In chromostereoscopic rendering 2D images are produced that encode depth in color. While the natural chromostereopsis of our human visual system is rather low, it can be enhanced via ChromaDepth R glasses, which induce chromatic aberrations in one Figure 2: Chromostereopsis can be due to: (a) longitunal chro- eye by refracting light of different wavelengths differently, hereby matic aberration, focus of blue shifts forward with respect to red, offsetting the projected position slightly in one eye. Although, it or (b) transverse chromatic aberration, blue shifts further toward might seem natural to map depth linearly to hue, which was also the the nasal part of the retina than red. (c) Shift in position leads to a basis of previous solutions, we demonstrate that such a mapping re- depth impression. duces the stereoscopic effect when using standard trichromatic dis- plays or printing systems. We propose an algorithm, which enables an improved stereoscopic experience with reduced artifacts.
    [Show full text]
  • Two Eyes See More Than One Human Beings Have Two Eyes Located About 6 Cm (About 2.4 In.) Apart
    ivi act ty 2 TTwowo EyesEyes SeeSee MoreMore ThanThan OneOne OBJECTIVES 1 straw 1 card, index (cut in half widthwise) Students discover how having two eyes helps us see in three dimensions and 3 pennies* increases our field of vision. 1 ruler, metric* The students For the class discover that each eye sees objects from a 1 roll string slightly different viewpoint to give us depth 1 roll tape, masking perception 1 pair scissors* observe that depth perception decreases *provided by the teacher with the use of just one eye measure their field of vision observe that their field of vision decreases PREPARATION with the use of just one eye Session I 1 Make a copy of Activity Sheet 2, Part A, for each student. SCHEDULE 2 Each team of two will need a metric ruler, Session I About 30 minutes three paper cups, and three pennies. Session II About 40 minutes Students will either close or cover their eyes, or you may use blindfolds. (Students should use their own blindfold—a bandanna or long strip VOCABULARY of cloth brought from home and stored in their science journals for use—in this depth perception and other activities.) field of vision peripheral vision Session II 1 Make a copy of Activity Sheet 2, Part B, for each student. MATERIALS 2 For each team, cut a length of string 50 cm (about 20 in.) long. Cut enough index For each student cards in half (widthwise) to give each team 1 Activity Sheet 2, Parts A and B half a card. Snip the corners of the cards to eliminate sharp edges.
    [Show full text]
  • Motion-In-Depth Perception and Prey Capture in the Praying Mantis Sphodromantis Lineola
    © 2019. Published by The Company of Biologists Ltd | Journal of Experimental Biology (2019) 222, jeb198614. doi:10.1242/jeb.198614 RESEARCH ARTICLE Motion-in-depth perception and prey capture in the praying mantis Sphodromantis lineola Vivek Nityananda1,*, Coline Joubier1,2, Jerry Tan1, Ghaith Tarawneh1 and Jenny C. A. Read1 ABSTRACT prey as they come near. Just as with depth perception, several cues Perceiving motion-in-depth is essential to detecting approaching could contribute to the perception of motion-in-depth. or receding objects, predators and prey. This can be achieved Two of the motion-in-depth cues that have received the most using several cues, including binocular stereoscopic cues such attention in humans are binocular: changing disparity and as changing disparity and interocular velocity differences, and interocular velocity differences (IOVDs) (Cormack et al., 2017). monocular cues such as looming. Although these have been Stereoscopic disparity refers to the difference in the position of an studied in detail in humans, only looming responses have been well object as seen by the two eyes. This disparity reflects the distance to characterized in insects and we know nothing about the role of an object. Thus as an object approaches, the disparity between the stereoscopic cues and how they might interact with looming cues. We two views changes. This changing disparity cue suffices to create a used our 3D insect cinema in a series of experiments to investigate perception of motion-in-depth for human observers, even in the the role of the stereoscopic cues mentioned above, as well as absence of other cues (Cumming and Parker, 1994).
    [Show full text]
  • From Stereogram to Surface: How the Brain Sees the World in Depth
    Spatial Vision, Vol. 22, No. 1, pp. 45–82 (2009) © Koninklijke Brill NV, Leiden, 2009. Also available online - www.brill.nl/sv From stereogram to surface: how the brain sees the world in depth LIANG FANG and STEPHEN GROSSBERG ∗ Department of Cognitive and Neural Systems, Center for Adaptive Systems and Center of Excellence for Learning in Education, Science, and Technology, Boston University, 677 Beacon St., Boston, MA 02215, USA Received 19 April 2007; accepted 31 August 2007 Abstract—When we look at a scene, how do we consciously see surfaces infused with lightness and color at the correct depths? Random-Dot Stereograms (RDS) probe how binocular disparity between the two eyes can generate such conscious surface percepts. Dense RDS do so despite the fact that they include multiple false binocular matches. Sparse stereograms do so even across large contrast-free regions with no binocular matches. Stereograms that define occluding and occluded surfaces lead to surface percepts wherein partially occluded textured surfaces are completed behind occluding textured surfaces at a spatial scale much larger than that of the texture elements themselves. Earlier models suggest how the brain detects binocular disparity, but not how RDS generate conscious percepts of 3D surfaces. This article proposes a neural network model that predicts how the layered circuits of visual cortex generate these 3D surface percepts using interactions between visual boundary and surface representations that obey complementary computational rules. The model clarifies how interactions between layers 4, 3B and 2/3A in V1 and V2 contribute to stereopsis, and proposes how 3D perceptual grouping laws in V2 interact with 3D surface filling-in operations in V1, V2 and V4 to generate 3D surface percepts in which figures are separated from their backgrounds.
    [Show full text]
  • Stereo Depth Perception
    Stereo Optical Company Vision Tester Slide Package: Driver Rehabilitation Slide Package For complete vision screening needs. Ideal for School, Athletic Physicals, Employment Physicals, Driver Licensing Exams, D.O.T., and General Check-Ups. Slide # 1: 2000-003 Slide # 2: 2000-010 Slide # 3: 2000-007 Slide # 4: 2000-012 Slide # 5: 2000-025 Slide # 6: 2000-013 Slide # 7: 2000-024 Slide # 8: 2000-087 Slide # 9: 2000-004 Slide #10: 2000-065 Slide #11: 2000-148 Slide #12: 2000-149 8600 W. Catalpa Ave., Suite 703 I Chicago, IL 60656 USA I 1.773.867.0380 I Toll-free 1.800.344.9500 Fax: 1.773.867.0388 I Email: [email protected] I www.StereoOptical.com P/N 56219 Slide 2000-003 “FAR” VISUAL ACUITY LETTERS 1. Dial at 1 (Yellow) indicator. 2. Far Point switch illuminated 3. Right and Left eye switches illuminated L R QUESTION: How many columns of letters do you see? SCORING: The answer is 3. Ask the subject to read line 5 completely. If this is correct, proceed to line 6, if correct proceed to line 7, if correct, the subject has 20/20 vision or better, at FAR point for the right eye, left eye and both eyes together. The correct scoring key is printed on the record form. Two or more letters incorrectly identified on any line (3 to 7) is considered a FAIL for that acuity level IN THAT COLUMN. A different acuity reading for each column is possible. The center column is critical because this is the binocular acuity test, while the right and left columns are monocular acuity tests.
    [Show full text]
  • Construction of Autostereograms Taking Into Account Object Colors and Its Applications for Steganography
    Construction of Autostereograms Taking into Account Object Colors and its Applications for Steganography Yusuke Tsuda Yonghao Yue Tomoyuki Nishita The University of Tokyo ftsuday,yonghao,[email protected] Abstract Information on appearances of three-dimensional ob- jects are transmitted via the Internet, and displaying objects q plays an important role in a lot of areas such as movies and video games. An autostereogram is one of the ways to represent three- dimensional objects taking advantage of binocular paral- (a) SS-relation (b) OS-relation lax, by which depths of objects are perceived. In previous Figure 1. Relations on an autostereogram. methods, the colors of objects were ignored when construct- E and E indicate left and right eyes, re- ing autostereogram images. In this paper, we propose a L R spectively. (a): A pair of points on the screen method to construct autostereogram images taking into ac- and a single point on the object correspond count color variation of objects, such as shading. with each other. (b): A pair of points on the We also propose a technique to embed color informa- object and a single point on the screen cor- tion in autostereogram images. Using our technique, au- respond with each other. tostereogram images shown on a display change, and view- ers can enjoy perceiving three-dimensional object and its colors in two stages. Color information is embedded in we propose an approach to take into account color variation a monochrome autostereogram image, and colors appear of objects, such as shading. Our approach assign slightly when the correct color table is used.
    [Show full text]
  • Stereoscopic Depth Perception at Far Viewing Distances
    Perception & Psychophysics 1984.35 (5). 423-428 Stereoscopic depth perception at far viewing distances ROBERT H. CORMACK New Mexico Institute ofMining and Technology, Socorro, New Mexico There has been some controversy as to whether stereoscopic depth constancy can operate at distances greater than a few meters. In the experiments reported here, observers judged the apparent depth in stereoscopic afterimages with crossed retinal disparity as fixation distance was changed. Apparent depth increased with increased fixation distance in accord with the geometry of retinal disparity. The results are discussed with reference to the perception of ap­ parent distance. Stereoscopic depth constancy refers to the veridical vergence (Foley, 1967). Furthermore, O'Leary and perception of depth mediated by retinal disparity as Wallach (1980) provide evidence that perspective and viewing distance is varied. Since a given depth inter­ familiar size can serve as cues for depth constancy. val produces less disparity at greater distances, stereo­ Given these findings, it is interesting that studies scopic depth constancy requires that retinal disparity ofstereopsis at distances greater than 2 m have found be rescaled as absolute distance changes. With sym­ a failure of depth constancy (Ono & Comerford, metrical convergence, and for objects within a few 1977). It is not at all clear why this should be. Naive, degrees of the line of sight, the depth (d) signaled by everyday viewing reveals no obvious departures from a given disparity can be calculated by the equation: constancy. Stereophotographs are capable of repre­ senting large depths at great apparent distances. Cues d = (.51(TAN(ATN(D/.51)+ .5R)))-D, (1) known to affect depth constancy such as linear per­ spective and familiar size could in principle provide where I is the interpupillary distance, R is the retinal the necessary information.
    [Show full text]
  • Abnormal Depth Perception from Motion Parallax in Amblyopic Observers
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Elsevier - Publisher Connector Vision Research 39 (1999) 1407–1413 Abnormal depth perception from motion parallax in amblyopic observers Angela M. Thompson, Mark Nawrot * Department of Psychology, North Dakota State Uni6ersity, Fargo, ND 58105, USA Received 4 February 1998; received in revised form 4 August 1998 Abstract Many similarities exist between the perception of depth from binocular stereopsis and that from motion parallax. Moreover, Rogers (1984, cited in, Howard, I. P., & Rogers, B. J. (1995). Binocular 6ision and stereopsis. Oxford Claridon, New York.) suggests a relationship between an observer’s ability to use disparity information and motion parallax information in a depth perception task. To more closely investigate this relationship, depth perception was studied in normal observers and amblyopic observers with poor stereo vision. As expected, amblyopic observers performed much worse than normal observers on depth discriminations requiring use of binocular disparity. However, amblyopic observers also performed much worse than normal observers on depth discriminations based on motion parallax. This result provides supporting evidence for a psychoanatomical link between the perception of depth from motion and the perception of depth from binocular disparity. © 1999 Elsevier Science Ltd. All rights reserved. Keywords: Motion parallax; Depth perception; Amblyopia 1. Introduction parallax is an ineffectual cue to depth (Epstein & Park, 1964; Gogel, 1977). However, using random-dot stim- Depth perception is of undeniable importance for uli, Rogers and Graham (1979) have convincingly quick and effortless locomotion through a cluttered and demonstrated that observers can accurately perceive hazard-filled environment.
    [Show full text]