Parallel Projection Modelling for Linear Array Scanner Scenes

Total Page:16

File Type:pdf, Size:1020Kb

Parallel Projection Modelling for Linear Array Scanner Scenes PARALLEL PROJECTION MODELLING FOR LINEAR ARRAY SCANNER SCENES M. Morgan a, K. Kim b, S. Jeong b, A. Habib a a Department of Geomatics Engineering, University of Calgary, Calgary, 2500 University Drive NW, Calgary, AB, T2N 1N4, Canada - ([email protected], [email protected]) b Electronics and Telecommunications Research Institute (ETRI), 161 Gajeong-Dong, Yuseong-Gu, Daejeon, 305-350, Korea – (kokim, soo) @etri.re.kr PS WG III/1: Sensor Pose Estimation KEY WORDS: Photogrammetry, Analysis, Modelling, Method, Pushbroom, Sensor, Stereoscopic, Value-added ABSTRACT: Digital frame cameras with resolution and ground coverage comparable to those associated with analogue aerial cameras are not yet available. Therefore, linear array scanners have been introduced on aerial and space borne platforms to overcome these drawbacks. Linear array scanner scenes are generated by stitching together successively captured one-dimensional images as the scanner moves. Rigorous modelling of these scenes is only possible if the internal and external characteristics of the imaging system (interior orientation parameters and exterior orientation parameters, respectively) are available. This is not usually the case (e.g., providers of IKONOS scenes do not furnish such information). Therefore, in order to obtain the parameters associated with the rigorous model, indirect estimation has to be performed. Space scenes with narrow angular field of view can lead to over-parameterization in indirect methods if the rigorous model is adopted. Earlier research has established that parallel projection can be used as an alternative/approximate model to express the imaging process of high altitude linear array scanners with narrow angular field of view. The parallel projection is attractive since it involves few parameters, which can be determined using limited number of ground control points. Moreover, the parallel projection model requires neither the interior nor the exterior orientation parameters of the imaging system. This paper outlines different parallel projection alternatives (linear and nonlinear). In addition, forward and backward transformations between these parameters are introduced. The paper also establishes the mathematical relationship between the navigation data, if available, and the parallel projection parameters. Finally, experimental results using synthetic data prove the suitability of parallel projection for modelling linear array scanners and verify the developed mathematical transformations. 1. INTRODUCTION Section 3. But first, background information regarding linear array scanner scenes is presented in Section 2. In cases where The limited number of pixels in 2-D digital images that are scanner navigation data are available, Section 4 sets up their captured by digital frame cameras limits their use in large scale relationship to the parallel projection model. Due to the mapping applications. On the other hand, scenes captured from complexity of the derivation of the developed transformations, linear scanners (also called pushbroom scanners or line Section 5 aims at verifying them by using synthetic data. cameras) have been introduced for their great potential of Finally, Section 6 includes the conclusions and generating ortho-photos and updating map databases (Wang, recommendations for future work. 1999). The linear scanners with up-to one-meter resolution from commercial satellites could bring more benefits and even a challenge to traditional topographic mapping with aerial images 2. BACKGROUND (Fritz, 1995). 2.1 Motivations for using Linear Array Scanners Careful sensor modelling has to be adapted in order to achieve the highest potential accuracy. Rigorous modelling describes the Two-dimensional digital cameras capture the data using a two- scene formation as it actually happens, and it has been adopted dimensional CCD array. However, the limited number of pixels in a variety of applications (Lee and Habib, 2002; Habib et al., in current digital imaging systems hinders their application 2001; Lee et al., 2000; Wang, 1999; Habib and Beshah, 1998; towards extended large scale mapping functions in comparison McGlone and Mikhail, 1981; Ethridge, 1977). with scanned analogue photographs. Increasing the principal distance of the 2-D digital cameras will increase the ground Alternatively, other approximate models exist such as rational resolution, but will decrease the ground coverage. On the other function model, RFM, direct linear transformation, DLT, self- hand, decreasing the principal distance will increase the ground calibrating DLT and parallel projection (Fraser et al., 2001; Tao coverage at the expense of ground resolution. and Hu, 2001; Dowman and Dolloff, 2000; Ono et al., 1999; Wang, 1999; Abdel-Aziz and Karara, 1971). Selection of any One-dimensional digital cameras (linear array scanners) can be approximate model, as an alternative, has to be done based on used to obtain large ground coverage and maintain a ground the analysis of the achieved accuracy. Among these models, resolution comparable with scanned analogue photographs. parallel projection is selected for the analysis. The rationale and However, they capture only a one-dimensional image (narrow the pre-requisites behind its selection are discussed as well as its strip) per snap shot. Ground coverage is achieved by moving the linear and non-linear mathematical forms are presented in scanner (airborne or space borne) and capturing more 1-D images. The scene of an area of interest is obtained by stitching as well as to increase the geometric strength of the together the resulting 1-D images. It is important to note that bundle adjustment. every 1-D image is associated with one-exposure station. Therefore, each 1-D image will have different Exterior EOP are either directly available from navigation units such as Orientation Parameters (EOP). A clear distinction is made GPS/INS mounted on the platform, or indirectly estimated using between the two terms “scene” and “image” throughout the ground control in bundle adjustment (Habib and Beshah, 1998; analysis of linear array scanners, Figure 1. Habib et al., 2001; Lee and Habib, 2002). For indirect estimation of the polynomial coefficients using Ground Control An image is defined as the recorded sensory data associated Points (GCP), instability of the bundle adjustment exists, with one exposure station. In the case of a frame image, it especially for space-based scenes (Wang, 1999; Fraser et al., contains only one exposure station, and consequently it is one 2001). This is attributed to the narrow Angular Field of View complete image. In the case of a linear array scanner, there are (AFOV) of space scenes, which results in very narrow bundles many 1-D images, each associated with a different exposure in the adjustment procedures. For this reason, a different model, station. The mathematical model that relates a point in the parallel projection, will be sought for the analysis in Section 3. object space and its corresponding point in the image space is the collinearity equations, which uses EOP of the appropriate image (in which the point appears). 3. PARALLEL PROJECTION In contrast, a scene is the recorded sensory data associated with 3.1 Motivations one (as in frame images) or more exposure stations (as in linear array scanners) that covers near continuous object space in a The motivations for selecting the parallel projection model to short single trip of the sensor. Therefore, in frame images, the approximate the rigorous model are summarized as follows: image and scene are identical terms, while in linear array scanners, the scene is an array of consecutive 1-D images. • Many space scanners have narrow AFOV – e.g., it is less than 1º for IKONOS scenes. For narrow AFOV, y y y y 1 2 i n the perspective light rays become closer to being parallel. x1 x2 xi xn • Space scenes are acquired within short time – e.g., it (a) is about one second for IKONOS scenes. Therefore, scanners can be assumed to have same attitude during y scene capturing. As a result, the planes, containing the images and their perspective centres, are parallel to each other. Column # i • For scenes captured in very short time, scanners can 1 2 i n or time be assumed to move with constant velocity. In this (b) case, the scanner travels equal distances in equal time Figure 1. A sequence of 1-D images (a) constituting a scene (b) intervals. As a result, same object distances are mapped into equal scene distances. 2.2 Rigorous Modelling of Linear Array Scanners Therefore, many space scenes, such as IKONOS, can be Rigorous (exact) modelling of linear array scanners describes assumed to comply with parallel projection. the actual geometric formation of the scenes at the time of photography. That involves the Interior Orientation Parameters 3.2 Forms of Parallel Projection Model (IOP) of the scanner and the EOP of each image in the scene. Representation of EOP, adopted by different researchers (Lee Figure 2 depicts a scene captured according to parallel and Habib, 2002; Lee et al., 2000; Wang, 1999; McGlone and projection. Scene parallel projection parameters include: two Mikhail, 198; Ethridge, 19771), includes: components of the unit projection vector (L, M); orientation angles of the scene (ω, ϕ, κ); two shift values (∆x, ∆y); and • Polynomial functions - This is motivated by the fact scale factor (s). Utilizing these parameters, the relationship that EOP do not abruptly change their values between between an object space point P(X, Y, Z), and the corresponding consecutive images in the scene, especially for space- scene point p(x, y’) can be expressed as: based scenes. • Piecewise polynomial functions - This option is ⎡ x ⎤ ⎡ L ⎤ ⎡ X ⎤ ⎡∆x⎤ (1) preferable if the scene time is large, and the variations ⎢ y'⎥ = s.λ.R T ⎢M ⎥ + s.R T ⎢Y ⎥ + ⎢∆y⎥ of EOP do not comply with one set of polynomial ⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥ functions. ⎣⎢ 0 ⎦⎥ ⎣⎢ N ⎦⎥ ⎣⎢ Z ⎦⎥ ⎣⎢ 0 ⎦⎥ • Orientation images – In this case, EOP of selected images (called orientation images) within the scene where: are explicitly dealt with. EOP of other images are R is the rotation matrix between the object and scene functions of those of the closest orientation images.
Recommended publications
  • CS 543: Computer Graphics Lecture 7 (Part I): Projection Emmanuel
    CS 543: Computer Graphics Lecture 7 (Part I): Projection Emmanuel Agu 3D Viewing and View Volume n Recall: 3D viewing set up Projection Transformation n View volume can have different shapes (different looks) n Different types of projection: parallel, perspective, orthographic, etc n Important to control n Projection type: perspective or orthographic, etc. n Field of view and image aspect ratio n Near and far clipping planes Perspective Projection n Similar to real world n Characterized by object foreshortening n Objects appear larger if they are closer to camera n Need: n Projection center n Projection plane n Projection: Connecting the object to the projection center camera projection plane Projection? Projectors Object in 3 space Projected image VRP COP Orthographic Projection n No foreshortening effect – distance from camera does not matter n The projection center is at infinite n Projection calculation – just drop z coordinates Field of View n Determine how much of the world is taken into the picture n Larger field of view = smaller object projection size center of projection field of view (view angle) y y z q z x Near and Far Clipping Planes n Only objects between near and far planes are drawn n Near plane + far plane + field of view = Viewing Frustum Near plane Far plane y z x Viewing Frustrum n 3D counterpart of 2D world clip window n Objects outside the frustum are clipped Near plane Far plane y z x Viewing Frustum Projection Transformation n In OpenGL: n Set the matrix mode to GL_PROJECTION n Perspective projection: use • gluPerspective(fovy,
    [Show full text]
  • Viewing in 3D
    Viewing in 3D Viewing in 3D Foley & Van Dam, Chapter 6 • Transformation Pipeline • Viewing Plane • Viewing Coordinate System • Projections • Orthographic • Perspective OpenGL Transformation Pipeline Viewing Coordinate System Homogeneous coordinates in World System zw world yw ModelViewModelView Matrix Matrix xw Tractor Viewing System Viewer Coordinates System ProjectionProjection Matrix Matrix Clip y Coordinates v Front- xv ClippingClipping Wheel System P0 zv ViewportViewport Transformation Transformation ne pla ing Window Coordinates View Specifying the Viewing Coordinates Specifying the Viewing Coordinates • Viewing Coordinates system, [xv, yv, zv], describes 3D objects with respect to a viewer zw y v P v xv •A viewing plane (projection plane) is set up N P0 zv perpendicular to zv and aligned with (xv,yv) yw xw ne pla ing • In order to specify a viewing plane we have View to specify: •P0=(x0,y0,z0) is the point where a camera is located •a vector N normal to the plane • P is a point to look-at •N=(P-P)/|P -P| is the view-plane normal vector •a viewing-up vector V 0 0 •V=zw is the view up vector, whose projection onto • a point on the viewing plane the view-plane is directed up Viewing Coordinate System Projections V u N z N ; x ; y z u x • Viewing 3D objects on a 2D display requires a v v V u N v v v mapping from 3D to 2D • The transformation M, from world-coordinate into viewing-coordinates is: • A projection is formed by the intersection of certain lines (projectors) with the view plane 1 2 3 ª x v x v x v 0 º ª 1 0 0 x 0 º « » «
    [Show full text]
  • Implementation of Projections
    CS488 Implementation of projections Luc RENAMBOT 1 3D Graphics • Convert a set of polygons in a 3D world into an image on a 2D screen • After theoretical view • Implementation 2 Transformations P(X,Y,Z) 3D Object Coordinates Modeling Transformation 3D World Coordinates Viewing Transformation 3D Camera Coordinates Projection Transformation 2D Screen Coordinates Window-to-Viewport Transformation 2D Image Coordinates P’(X’,Y’) 3 3D Rendering Pipeline 3D Geometric Primitives Modeling Transform into 3D world coordinate system Transformation Lighting Illuminate according to lighting and reflectance Viewing Transform into 3D camera coordinate system Transformation Projection Transform into 2D camera coordinate system Transformation Clipping Clip primitives outside camera’s view Scan Draw pixels (including texturing, hidden surface, etc.) Conversion Image 4 Orthographic Projection 5 Perspective Projection B F 6 Viewing Reference Coordinate system 7 Projection Reference Point Projection Reference Point (PRP) Center of Window (CW) View Reference Point (VRP) View-Plane Normal (VPN) 8 Implementation • Lots of Matrices • Orthographic matrix • Perspective matrix • 3D World → Normalize to the canonical view volume → Clip against canonical view volume → Project onto projection plane → Translate into viewport 9 Canonical View Volumes • Used because easy to clip against and calculate intersections • Strategies: convert view volumes into “easy” canonical view volumes • Transformations called Npar and Nper 10 Parallel Canonical Volume X or Y Defined by 6 planes
    [Show full text]
  • EX NIHILO – Dahlgren 1
    EX NIHILO – Dahlgren 1 EX NIHILO: A STUDY OF CREATIVITY AND INTERDISCIPLINARY THOUGHT-SYMMETRY IN THE ARTS AND SCIENCES By DAVID F. DAHLGREN Integrated Studies Project submitted to Dr. Patricia Hughes-Fuller in partial fulfillment of the requirements for the degree of Master of Arts – Integrated Studies Athabasca, Alberta August, 2008 EX NIHILO – Dahlgren 2 Waterfall by M. C. Escher EX NIHILO – Dahlgren 3 Contents Page LIST OF ILLUSTRATIONS 4 INTRODUCTION 6 FORMS OF SIMILARITY 8 Surface Connections 9 Mechanistic or Syntagmatic Structure 9 Organic or Paradigmatic Structure 12 Melding Mechanical and Organic Structure 14 FORMS OF FEELING 16 Generative Idea 16 Traits 16 Background Control 17 Simulacrum Effect and Aura 18 The Science of Creativity 19 FORMS OF ART IN SCIENTIFIC THOUGHT 21 Interdisciplinary Concept Similarities 21 Concept Glossary 23 Art as an Aid to Communicating Concepts 27 Interdisciplinary Concept Translation 30 Literature to Science 30 Music to Science 33 Art to Science 35 Reversing the Process 38 Thought Energy 39 FORMS OF THOUGHT ENERGY 41 Zero Point Energy 41 Schools of Fish – Flocks of Birds 41 Encapsulating Aura in Language 42 Encapsulating Aura in Art Forms 50 FORMS OF INNER SPACE 53 Shapes of Sound 53 Soundscapes 54 Musical Topography 57 Drawing Inner Space 58 Exploring Inner Space 66 SUMMARY 70 REFERENCES 71 APPENDICES 78 EX NIHILO – Dahlgren 4 LIST OF ILLUSTRATIONS Page Fig. 1 - Hofstadter’s Lettering 8 Fig. 2 - Stravinsky by Picasso 9 Fig. 3 - Symphony No. 40 in G minor by Mozart 10 Fig. 4 - Bird Pattern – Alhambra palace 10 Fig. 5 - A Tree Graph of the Creative Process 11 Fig.
    [Show full text]
  • CS 4204 Computer Graphics 3D Views and Projection
    CS 4204 Computer Graphics 3D views and projection Adapted from notes by Yong Cao 1 Overview of 3D rendering Modeling: * Topic we’ve already discussed • *Define object in local coordinates • *Place object in world coordinates (modeling transformation) Viewing: • Define camera parameters • Find object location in camera coordinates (viewing transformation) Projection: project object to the viewplane Clipping: clip object to the view volume *Viewport transformation *Rasterization: rasterize object Simple teapot demo 3D rendering pipeline Vertices as input Series of operations/transformations to obtain 2D vertices in screen coordinates These can then be rasterized 3D rendering pipeline We’ve already discussed: • Viewport transformation • 3D modeling transformations We’ll talk about remaining topics in reverse order: • 3D clipping (simple extension of 2D clipping) • 3D projection • 3D viewing Clipping: 3D Cohen-Sutherland Use 6-bit outcodes When needed, clip line segment against planes Viewing and Projection Camera Analogy: 1. Set up your tripod and point the camera at the scene (viewing transformation). 2. Arrange the scene to be photographed into the desired composition (modeling transformation). 3. Choose a camera lens or adjust the zoom (projection transformation). 4. Determine how large you want the final photograph to be - for example, you might want it enlarged (viewport transformation). Projection transformations Introduction to Projection Transformations Mapping: f : Rn Rm Projection: n > m Planar Projection: Projection on a plane.
    [Show full text]
  • Shortcuts Guide
    Shortcuts Guide One Key Shortcuts Toggles and Screen Management Hot Keys A–Z Printable Keyboard Stickers ONE KEY SHORTCUTS [SEE PRINTABLE KEYBOARD STICKERS ON PAGE 11] mode mode mode mode Help text screen object 3DOsnap Isoplane Dynamic UCS grid ortho snap polar object dynamic mode mode Display Toggle Toggle snap Toggle Toggle Toggle Toggle Toggle Toggle Toggle Toggle snap tracking Toggle input PrtScn ScrLK Pause Esc F1 F2 F3 F4 F5 F6 F7 F8 F9 F10 F11 F12 SysRq Break ~ ! @ # $ % ^ & * ( ) — + Backspace Home End ` 1 2 3 4 5 6 7 8 9 0 - = { } | Tab Q W E R T Y U I O P Insert Page QSAVE WBLOCK ERASE REDRAW MTEXT INSERT OFFSET PAN [ ] \ Up : “ Caps Lock A S D F G H J K L Enter Delete Page ARC STRETCH DIMSTYLE FILLET GROUP HATCH JOIN LINE ; ‘ Down < > ? Shift Z X C V B N M Shift ZOOM EXPLODE CIRCLE VIEW BLOCK MOVE , . / Ctrl Start Alt Alt Ctrl Q QSAVE / Saves the current drawing. C CIRCLE / Creates a circle. H HATCH / Fills an enclosed area or selected objects with a hatch pattern, solid fill, or A ARC / Creates an arc. R REDRAW / Refreshes the display gradient fill. in the current viewport. Z ZOOM / Increases or decreases the J JOIN / Joins similar objects to form magnification of the view in the F FILLET / Rounds and fillets the edges a single, unbroken object. current viewport. of objects. M MOVE / Moves objects a specified W WBLOCK / Writes objects or V VIEW / Saves and restores named distance in a specified direction. a block to a new drawing file.
    [Show full text]
  • Viewing and Projection Viewing and Projection
    Viewing and Projection The topics • Interior parameters • Projection type • Field of view • Clipping • Frustum… • Exterior parameters • Camera position • Camera orientation Transformation Pipeline Local coordinate Local‐>World World coordinate ModelView World‐>Eye Matrix Eye coordinate Projection Matrix Clip coordina te others Screen coordinate Projection • The projection transforms a point from a high‐ dimensional space to a low‐dimensional space. • In 3D, the projection means mapping a 3D point onto a 2D projection plane (or called image plane). • There are two basic projection types: • Parallel: orthographic, oblique • Perspective Orthographic Projection Image Plane Direction of Projection z-axis z=k x 1000 x y 0100 y k 000k z 1 0001 1 Orthographic Projection Oblique Projection Image Plane Direction of Projection Properties of Parallel Projection • Definition: projection directions are parallel. • Doesn’t look real. • Can preserve parallel lines Projection PlllParallel in 3D PlllParallel in 2D Properties of Parallel Projection • Definition: projection directions are parallel. • Doesn’t look real. • Can preserve parallel lines • Can preserve ratios t ' t Projection s s :t s' :t ' s' Properties of Parallel Projection • Definition: projection directions are parallel. • Doesn’t look real. • Can preserve parallel lines • Can preserve ratios • CANNOT preserve angles Projection Properties of Parallel Projection • Definition: projection directions are parallel. • Doesn’t look real. • Can preserve parallel
    [Show full text]
  • Image Formation • Projection Geometry • Radiometry (Image
    Image Formation • Projection Geometry • Radiometry (Image Brightness) - to be discussed later in SFS. Image Formation 1 Pinhole Camera (source: A Guided tour of computer vision/Vic Nalwa) Image Formation 2 Perspective Projection (source: A Guided tour of computer vision/Vic Nalwa) Image Formation 3 Perspective Projection Image Formation 4 Some Observations/questions • Note that under perspective projection, straight- lines in 3-D project as straight lines in the 2-D image plane. Can you prove this analytically? – What is the shape of the image of a sphere? – What is the shape of the image of a circular disk? Assume that the disk lies in a plane that is tilted with respect to the image plane. • What would be the image of a set of parallel lines – Do they remain parallel in the image plane? Image Formation 5 Note: Equation for a line in 3-D (and in 2-D) Line in 3-D: Line in 2-D By using the projective geometry equations, it is easy to show that a line in 3-D projects as a line in 2-D. Image Formation 6 Vanishing Point • Vanishing point of a straight line under perspective projection is that point in the image beyond which the projection of the straight line can not extend. – I.e., if the straight line were infinitely long in space, the line would appear to vanish at its vanishing point in the image. – The vanishing point of a line depends ONLY on its orientation is space, and not on its position. – Thus, parallel lines in space appear to meet at their vanishing point in image.
    [Show full text]
  • A General-Purpose Animation System for 4D Justin Alain Jensen Brigham Young University
    Brigham Young University BYU ScholarsArchive All Theses and Dissertations 2017-08-01 A General-Purpose Animation System for 4D Justin Alain Jensen Brigham Young University Follow this and additional works at: https://scholarsarchive.byu.edu/etd Part of the Computer Sciences Commons BYU ScholarsArchive Citation Jensen, Justin Alain, "A General-Purpose Animation System for 4D" (2017). All Theses and Dissertations. 6968. https://scholarsarchive.byu.edu/etd/6968 This Thesis is brought to you for free and open access by BYU ScholarsArchive. It has been accepted for inclusion in All Theses and Dissertations by an authorized administrator of BYU ScholarsArchive. For more information, please contact [email protected], [email protected]. A General-Purpose Animation System for 4D Justin Alain Jensen A thesis submitted to the faculty of Brigham Young University in partial fulfillment of the requirements for the degree of Master of Science Robert P. Burton, Chair Parris K. Egbert Seth R. Holladay Department of Computer Science Brigham Young University Copyright c 2017 Justin Alain Jensen All Rights Reserved ABSTRACT A General-Purpose Animation System for 4D Justin Alain Jensen Department of Computer Science, BYU Master of Science Computer animation has been limited almost exclusively to 2D and 3D. The tools for 3D computer animation have been largely in place for decades and are well-understood. Existing tools for visualizing 4D geometry include minimal animation features. Few tools have been designed specifically for animation of higher-dimensional objects, phenomena, or spaces. None have been designed to be familiar to 3D animators. A general-purpose 4D animation system can be expected to facilitate more widespread understanding of 4D geometry and space, can become the basis for creating unique 3D visual effects, and may offer new insight into 3D animation concepts.
    [Show full text]
  • Chapter 2 – Part 2 Multiview Drawing Chapter Objectives
    Chapter 2 – Part 2 Multiview Drawing Chapter Objectives •Explain what multiview drawings are and their importance to the field of technical drawing. •Explain how views are chosen and aligned in a multiview drawing. •Visualize and interpret the multiviews of an object. •Describe projection planes •Describe normal, inclined and oblique surfaces •Describe line types •Describe line weights •Interpret the multiviews of graphic primitives •Describe orthographic projection including the miter line technique Supplemental Files •Describe the line types and line weights used in technical drawings as defined by the ASME Y14.2M standard. •Explain the difference between drawings created with First Angle and Third Angle projection techniques. •Use a miter line to project information between top and side views. •Create multiview sketches of objects including the correct placement and depiction of visible, hidden and center lines. Supplemental files are available for download inside the Chapter 2 folder of the book’s file downloads. Please see the inside front cover for further details. © Technical Drawing 101 with AutoCAD Smith & Ramirez – All rights reserved. 2.4 MULTIVIEW DRAWINGS Multiview drawing is a technique used by drafters and designers to depict a three-dimensional object (an object having height, width and depth) as a group of related two-dimensional (having only width and height, or width and depth) views. A person trained in interpreting multiview drawings, can visualize an object’s three-dimensional shape by studying the two-dimensional multiview drawings of the object. For example, Figure 2.15 provides a three-dimensional (3D) image of a school bus, and while a 3D view of the bus is very helpful in visualizing its overall shape, it doesn’t show the viewer all of the sides of the bus, or the true length, width, or height of the bus.
    [Show full text]
  • The Parallel Projection, As Flights of Fancy
    Portland State University PDXScholar Proceedings of the 18th National Conference on the Architecture Beginning Design Student 2002 The aP rallel Projection, as Flights of Fancy Mary Nixon University of Pennsylvania Let us know how access to this document benefits ouy . Follow this and additional works at: http://pdxscholar.library.pdx.edu/arch_design Part of the Architecture Commons, and the Art and Design Commons Recommended Citation Nixon, Mary, "The aP rallel Projection, as Flights of Fancy" (2002). Proceedings of the 18th National Conference on the Beginning Design Student. Paper 37. http://pdxscholar.library.pdx.edu/arch_design/37 This Article is brought to you for free and open access. It has been accepted for inclusion in Proceedings of the 18th National Conference on the Beginning Design Student by an authorized administrator of PDXScholar. For more information, please contact [email protected]. 1111 18th National Conference on the Beginning Design Student Portland. Oregon . 2002 The parallel projection, as flights of fancy Mary Nixon University of Pennsylvania The poet's eye, in a fine frenzy rolling, Doth glance from includes the families of axonometrics, isometrics (having iden­ heaven to earth, from earth to heaven; And as imagina­ tical standard measurement in all three axes), dimetrics (in tion bodies forth The forms of things unknown, the two axes), trimetrics (in none of the three axes), and obliques. poet's pen Tums them to shapes, and gives to airy noth­ They are all variations on the same theme. "Parallel lines 1 ing A local habitation and a name. remain parallel in parallel projections, while they converge to William Shakespeare, through Theseus, beautifully describes vanishing points in (central projections or) perspectives."4 this creative method of the poet.
    [Show full text]
  • CARTOGRAPHIC PORTRAYAL of TERRAIN in OBLIQUE PARALLEL PROJECTION Pyry Kettunen, Tapani Sarjakoski, L
    CARTOGRAPHIC PORTRAYAL OF TERRAIN IN OBLIQUE PARALLEL PROJECTION Pyry Kettunen, Tapani Sarjakoski, L. Tiina Sarjakoski, Juha Oksanen Finnish Geodetic Institute Department of Geoinformatics and Cartography P.O. Box 15 FI-02431 Masala Finland e-mails: [email protected] Abstract The visualization of three-dimensional spatial information as oblique views is widely regarded as beneficial for the perceiving of three-dimensional space. From the cartographic perspective, oblique views need to be carefully designed, in order to help the map reader to form a useful mental model of the terrain. In this study, 3D cartographic methods were applied to illustrate directions, distances and heights in oblique views. An oblique parallel projection was used for representing the view in a metrically homogeneous manner and for emphasizing the relative heights. The height was further visualized using directed lighting, hypsographic coloring and contours. In addition, an equilateral and equiangular grid was draped on the terrain model, for fixing the cardinal directions and for facilitating the estimation of distances and surface areas. These methods were combined and analyzed in a case study on a high-resolution DEM covering a national park area. The resulting oblique parallel views provide visual cues for the estimation of terrain dimensions, particularly heights. The results prepare a basis for further research on cartographic 3D visualization. Keywords: 3D cartography, oblique parallel projection, map grid, DEM 1 INTRODUCTION Outdoor leisure activities, such as forest hiking, have increasingly gained in popularity. These activities often benefit from navigational aids, such as route markings and maps. Recent technological developments have led to the successful implementation of personal navigation applications on mobile devices, which primarily focus on city life.
    [Show full text]