235-242.Pdf (449.2Kb)

Total Page:16

File Type:pdf, Size:1020Kb

235-242.Pdf (449.2Kb) Eurographics/ IEEE-VGTC Symposium on Visualization (2007) Ken Museth, Torsten Möller, and Anders Ynnerman (Editors) Hardware-accelerated Stippling of Surfaces derived from Medical Volume Data Alexandra Baer Christian Tietjen Ragnar Bade Bernhard Preim Department of Simulation and Graphics Otto-von-Guericke University of Magdeburg, Germany {abaer|tietjen|rbade|preim}@isg.cs.uni-magdeburg.de Abstract We present a fast hardware-accelerated stippling method which does not require any preprocessing for placing points on surfaces. The surfaces are automatically parameterized in order to apply stippling textures without ma- jor distortions. The mapping process is guided by a decomposition of the space in cubes. Seamless scaling with a constant density of points is realized by subdividing and summarizing cubes. Our mip-map technique enables ar- bitrarily scaling with one texture. Different shading tones and scales are facilitated by adhering to the constraints of tonal art maps. With our stippling technique, it is feasible to encode all scaling and brightness levels within one self-similar texture. Our method is applied to surfaces extracted from (segmented) medical volume data. The speed of the stippling process enables stippling for several complex objects simultaneously. We consider applica- tion scenarios in intervention planning (neck and liver surgery planning). In these scenarios, object recognition (shape perception) is supported by adding stippling to semi-transparently shaded objects which are displayed as context information. Categories and Subject Descriptors (according to ACM CCS): I.3.7 [Computer Graphics]: Three-Dimensional Graphics and Realism -Color, shading, shadowing, and texture 1. Introduction from breathing, pulsation or patient movement. Without a well-defined curvature field, hatching may be misleading by The effective visualization of complex anatomic surfaces is a emphasizing erroneous features of the surfaces. Stippling, challenging task since nested surfaces and complex geome- in principle, requires high-quality surface models. However, tries are involved. With conventional surface and volume since stippling does not produce visually striking primitives rendering, it is often not possible to convey object shapes and such as lines, this technique is better suited for visualizing depth relations precisely. This gave rise to sparse visual rep- surface models with moderate or average quality. Compared resentations, such as hatching and stippling [VKG05]. Saito to hatching, stippling is more adequate for objects which do and Takahashi [ST90] as well as Interrante and Kim [IK01] not exhibit ridges, valleys or other regions of high curvature. showed that depth perception of medical surface models can be effectively enhanced by hatching along curvature direc- Stippling is a rendering technique, where small dots are tions. Hatching requires high-quality smooth surface mod- used to convey shape and shading. The points’ density and els which are very difficult to create automatically based on their location on the surface affect the appearance of a stip- clinical patient data exhibiting noise and minor artifacts, e.g. pled object. While densely distributed dots achieve the im- © The Eurographics Association 2007. 236 Baer et. al / Hardware-accelerated Stippling of Surfaces derived from Medical Volume Data pression of dark shading, sparsely covered regions seem like even because the object moves over a sequence of frames. brighter shaded areas. The points are distributed randomly, This "shower door" effect hampers the use of this technique but the spacing among the points is regular without any lin- in animation sequences, in which particles should move ear or regular pattern. along with the object. We present an effective hardware-accelerated stippling In object space approaches, NPR primitives are associ- method. Compared with existing methods, our approach has ated with actual locations on the model’s surface to achieve the following advantages: frame-coherence. Meier [Mei96] used painterly strokes and Cornish et al. [CRL01] developed a view-dependent parti- 1. It is frame-coherent and therefore suitable for animations cle system in which the distribution and density of strokes is and interactive exploration. All operations for texture co- regulated by a hierarchical clustering algorithm. Similarly, ordinate generation, access and mapping are supported Pastor and Strothotte [PFS03] proposed a concept to gener- by graphics hardware. ate interactive stippling animations using graphics hardware. 2. Our multi-cube mapping algorithm minimizes texture They consider each vertex as a potential location of one stip- distortions and integrates different shading tones and pling point. The amount of stippling points, according to scales in one texture while satisfying the constraints of shading tones and scaling, will be achieved by subdivision Tonal Art Maps [PHWF01]. and simplification of the object geometry. As the amount of 3. The performance is independent of the number of stip- points depends on the number of vertices, time consuming pling points. Moreover, no preprocessing is necessary to preprocessing steps are necessary. The point distribution is realize a stippling rendering for any object. realized with a point hierarchy based on the polygonal mesh In this paper, we describe our stippling approach and and a point relaxation. present scenarios, in which stippling is applied for inter- Lu et al. [LME∗02] introduced an interactive volume vi- vention planning purposes. In contrast to medical education, sualization that simulates stippling drawings, thus combin- where the author may, in principle, work for a long time to ing traditional volume and NPR illustration techniques. Each create a few expressive visualizations, intervention planning volume is illustrated with an appropriate number of stip- is a routine task, which has to be accomplished very fast. pling points initially generated. Treavett and Chen [TC00] Therefore, automatic adjustment of parameters and render- and Sousa et al. [SESS05] apply NPR techniques to scien- ing speed are also considered. Stippling should be used for tific and medical datasets to improve comprehensibility of context objects which serve as orientation aids and not for medical education and surgical training examples. the focus objects, e. g. a tumor, which are immediately rel- evant for an intervention. Therefore, accuracy is not an im- Texture-based methods combine aspects of image- and portant issue. The major source of uncertainty for these visu- object-based approaches. Textures with stippling points or alizations is the segmentation process, not the stippling ap- hatching strokes are mapped to the surface of the in- proach. put model. The first stippled renditions were presented by Winkenbach and Salesin [WS94] for parametric surfaces us- ing randomness to distribute the points on the texture and 2. Related Work therefore on the surface of a model. They also developed prioritized stroke textures to apply a hatching pattern in ob- Image-, object- and texture-based methods are the main ap- ∗ proaches to illustrate objects with stippling or other render- ject space. Klein et al. [KLK 00] used Art Maps and Praun ing styles. Deussen et al. [DHvOS00] and Secord [Sec02] et al. [PHWF01] introduced Tonal Art Maps (TAMs) and produce high-quality stippling images based on greyscale lapped textures. TAMs are a sequence of hatch images rep- images as input. Both use distribution methods based on resenting different tones and mip-map levels. Strokes within Voronoi diagrams and relaxation to achieve a regular posi- the images are scaled to have an appropriate stroke size and tioning of stippling points that avoids nearly all visual pat- density at all resolutions. They are organized to maintain co- terns. While the presented images are visually attractive, herence across scale and tones. Yuan et al. [YNZC05] intro- these image-based approaches are restricted to single rendi- duced a 2D geometry-image processing method to generate tions. The stippling point distribution obtained in one frame stippling renditions. Points are directly rendered and there- is not guaranteed to correspond with the one obtained in the fore the frame rate is roughly inversely proportional to the next frame. Secord et al. [SHS02] propose a method where amount of generated stippling points. frame-coherence is pursued on the image plane, not in ob- ject space. Stippling points or hatching strokes (see Eissele 3. Theoretical Background and Overview et al. [EWE04]) are attached to the model’s surface. This ap- proach is not frame-coherent in object space and the non- Image space methods are normally fast, but since the points photorealistic rendering (NPR) primitives (e.g. points for are placed in image space, undesirable effects (e. g."shower stippling or strokes for hatching) will move across the ob- door" effect) may appear. Object space methods are frame ject’s surface during an animation by shading changes or coherent but commonly require time-consuming preprocess- © The Eurographics Association 2007. Baer et. al / Hardware-accelerated Stippling of Surfaces derived from Medical Volume Data 237 ing steps. Texture-based methods are as fast as image- • the texture design including the representation of various based techniques, frame-coherent like object-based ap- shading tones by the density of stippling points and proaches and furthermore predominantly geometry indepen- • the texture mapping process based on a parametrization
Recommended publications
  • Relief Texture Mapping
    Advanced Texturing Environment Mapping Environment Mapping ° reflections Environment Mapping ° orientation ° Environment Map View point Environment Mapping ° ° ° Environment Map View point Environment Mapping ° Can be an “effect” ° Usually means: “fake reflection” ° Can be a “technique” (i.e., GPU feature) ° Then it means: “2D texture indexed by a 3D orientation” ° Usually the index vector is the reflection vector ° But can be anything else that’s suitable! ° Increased importance for modern GI Environment Mapping ° Uses texture coordinate generation, multi-texturing, new texture targets… ° Main task ° Map all 3D orientations to a 2D texture ° Independent of application to reflections Sphere Cube Dual paraboloid Top top Top Left Front Right Left Front Right Back left front right Bottom Bottom Back bottom Cube Mapping ° OpenGL texture targets Top Left Front Right Back Bottom glTexImage2D( GL_TEXTURE_CUBE_MAP_POSITIVE_X , 0, GL_RGB8, w, h, 0, GL_RGB, GL_UNSIGNED_BYTE, face_px ); Cube Mapping ° Cube map accessed via vectors expressed as 3D texture coordinates (s, t, r) +t +s -r Cube Mapping ° 3D 2D projection done by hardware ° Highest magnitude component selects which cube face to use (e.g., -t) ° Divide other components by this, e.g.: s’ = s / -t r’ = r / -t ° (s’, r’) is in the range [-1, 1] ° remap to [0,1] and select a texel from selected face ° Still need to generate useful texture coordinates for reflections Cube Mapping ° Generate views of the environment ° One for each cube face ° 90° view frustum ° Use hardware render to texture
    [Show full text]
  • Generating Real Time Reflections by Ray Tracing Approximate Geometry
    Generating Real Time Reflections by Ray Tracing Approximate Geometry Master’s thesis in Interaction Design and Technologies JOHAN FREDRIKSSON, ADAM SCOTT Department of Interaction Design and Technologies CHALMERS UNIVERSITY OF TECHNOLOGY Gothenburg, Sweden 2016 Master’s thesis 2016:123 Generating Real Time Reflections by Ray Tracing Approximate Geometry JOHAN FREDRIKSSON, ADAM SCOTT Department of Computer Science Division of Interaction Design and Technologies Chalmers University of Technology Gothenburg, Sweden 2016 Generating Real Time Reflections by Ray Tracing Approximate Geometry JOHAN FREDRIKSSON, ADAM SCOTT © JOHAN FREDRIKSSON, ADAM SCOTT, 2016. Supervisor: Erik Sintorn, Computer Graphics Examiner: Staffan Björk, Interaction Design and Technologies Master’s Thesis 2016:123 Department of Computer Science Division of Interaction Design and Technologies Chalmers University of Technology SE-412 96 Gothenburg Cover: BVH encapsulating approximate geometry created in the Frostbite engine. Typeset in LATEX Gothenburg, Sweden 2016 iii Generating Real Time Reflections by Ray Tracing Approximated Geometry Johan Fredriksson, Adam Scott Department of Interaction Design and Technologies Chalmers University of Technology Abstract Rendering reflections in real time is important for realistic modern games. A com- mon way to provide reflections is by using environment mapping. This method is not an accurate way to calculate reflections and it consumes a lot of memory space. We propose a method of ray tracing approximate geometry in real time. By using approximate geometry it can be kept simple and the size of the bounding volume hierarchy will have a lower memory impact. The ray tracing is performed on the GPU and the bounding volumes are pre built on the CPU using surface area heuris- tics.
    [Show full text]
  • Brno University of Technology Rendering Of
    BRNO UNIVERSITY OF TECHNOLOGY VYSOKÉ UČENÍ TECHNICKÉ V BRNĚ FACULTY OF INFORMATION TECHNOLOGY FAKULTA INFORMAČNÍCH TECHNOLOGIÍ DEPARTMENT OF COMPUTER GRAPHICS AND MULTIMEDIA ÚSTAV POČÍTAČOVÉ GRAFIKY A MULTIMÉDIÍ RENDERING OF NONPLANAR MIRRORS ZOBRAZOVÁNÍ POKŘIVENÝCH ZRCADEL MASTER’S THESIS DIPLOMOVÁ PRÁCE AUTHOR Bc. MILOSLAV ČÍŽ AUTOR PRÁCE SUPERVISOR Ing. TOMÁŠ MILET VEDOUCÍ PRÁCE BRNO 2017 Abstract This work deals with the problem of accurately rendering mirror reflections on curved surfaces in real-time. While planar mirrors do not pose a problem in this area, non-planar surfaces are nowadays rendered mostly using environment mapping, which is a method of approximating the reflections well enough for the human eye. However, this approach may not be suitable for applications such as CAD systems. Accurate mirror reflections can be rendered with ray tracing methods, but not in real-time and therefore without offering interactivity. This work examines existing approaches to the problem and proposes a new algorithm for computing accurate mirror reflections in real-time using accelerated searching for intersections with depth profile stored in cubemap textures. This algorithm has been implemented using OpenGL and tested on different platforms. Abstrakt Tato práce se zabývá problémem přesného zobrazování zrcadlových odrazů na zakřiveném povrchu v reálném čase. Zatímco planární zrcadla nepředstavují v tomto ohledu problém, zakřivené povrchy se v dnešní době zobrazují především metodou environment mapping, která aproximuje reálné odrazy a nabízí výsledky uspokojivé pro lidské oko. Tento přístup však nemusí být vhodný např. v oblasti CAD systémů. Přesných zrcadlových odrazů se dá dosáhnout pomocí metod sledování paprsku, avšak ne v reálném čase a tudíž bez možnosti interaktivity.
    [Show full text]
  • 2.5D 3D Axis-Aligned Bounding Box (AABB) Albedo Alpha Blending
    This page is dedicated to describing the commonly used terms when working with CRYENGINE V. As a use case, you may find yourself asking questions like "What is a Brush", "What is a Component", or Overview | Sections | # | A | B | C | D | E "What is POM?", this page will highlight and provide descriptions for those types of questions and more. | F | G | H | I | J | L | M | N | O | P | Q | R Once you have an understanding of what each term means, links are provided for more documentation | S | T | V | Y | Z and guidance on working with them. 2.5D Used to describe raycasting engines, such as Doom or Duke Nukem. The maps are generally drawn as 2D bitmaps with height properties but when rendered gives the appearance of being 3D. 3D Refers to the ability to move in the X, Y and Z axis and also rotate around the X, Y, and Z axes. Axis-aligned Bounding Box A form of a bounding box where the box is aligned to the axis therefore only two points in space are (AABB) needed to define it. AABB’s are much faster to use, and take up less memory, but are very limited in the sense that they can only be aligned to the axis. Albedo An albedo is a physical measure of a material’s reflectivity, generally across the visible spectrum of light. An albedo texture simulates a surface albedo as opposed to explicitly defining a colour for it. Alpha Blending Assigning varying levels of translucency to graphical objects, allowing the creation of things such as glass, fog, and ghosts.
    [Show full text]
  • Real-Time Recursive Specular Reflections on Planar and Curved Surfaces Using Graphics Hardware
    REAL-TIME RECURSIVE SPECULAR REFLECTIONS ON PLANAR AND CURVED SURFACES USING GRAPHICS HARDWARE Kasper Høy Nielsen Niels Jørgen Christensen Informatics and Mathematical Modelling The Technical University of Denmark DK 2800 Lyngby, Denmark {khn,njc}@imm.dtu.dk ABSTRACT Real-time rendering of recursive reflections have previously been done using different techniques. How- ever, a fast unified approach for capturing recursive reflections on both planar and curved surfaces, as well as glossy reflections and interreflections between such primitives, have not been described. This paper describes a framework for efficient simulation of recursive specular reflections in scenes containing both planar and curved surfaces. We describe and compare two methods that utilize texture mapping and en- vironment mapping, while having reasonable memory requirements. The methods are texture-based to allow for the simulation of glossy reflections using image-filtering. We show that the methods can render recursive reflections in static and dynamic scenes in real-time on current consumer graphics hardware. The methods make it possible to obtain a realism close to ray traced images at interactive frame rates. Keywords: Real-time rendering, recursive specular reflections, texture, environment mapping. 1 INTRODUCTION mapped surface sees the exact same surroundings re- gardless of its position. Static precalculated environ- Mirror-like reflections are known from ray tracing, ment maps can be used to create the notion of reflec- where shiny objects seem to reflect the surrounding tion on shiny objects. However, for modelling reflec- environment. Rays are continuously traced through tions similar to ray tracing, reflections should capture the scene by following reflected rays recursively un- the actual synthetic environment as accurately as pos- til a certain depth is reached, thereby capturing both sible, while still being feasible for interactive render- reflections and interreflections.
    [Show full text]
  • Ray Tracing Tutorial
    Ray Tracing Tutorial by the Codermind team Contents: 1. Introduction : What is ray tracing ? 2. Part I : First rays 3. Part II : Phong, Blinn, supersampling, sRGB and exposure 4. Part III : Procedural textures, bump mapping, cube environment map 5. Part IV : Depth of field, Fresnel, blobs 2 This article is the foreword of a serie of article about ray tracing. It's probably a word that you may have heard without really knowing what it represents or without an idea of how to implement one using a programming language. This article and the following will try to fill that for you. Note that this introduction is covering generalities about ray tracing and will not enter into much details of our implementation. If you're interested about actual implementation, formulas and code please skip this and go to part one after this introduction. What is ray tracing ? Ray tracing is one of the numerous techniques that exist to render images with computers. The idea behind ray tracing is that physically correct images are composed by light and that light will usually come from a light source and bounce around as light rays (following a broken line path) in a scene before hitting our eyes or a camera. By being able to reproduce in computer simulation the path followed from a light source to our eye we would then be able to determine what our eye sees. Of course it's not as simple as it sounds. We need some method to follow these rays as the nature has an infinite amount of computation available but we have not.
    [Show full text]
  • (543) Lecture 7 (Part 2): Texturing Prof Emmanuel
    Computer Graphics (543) Lecture 7 (Part 2): Texturing Prof Emmanuel Agu Computer Science Dept. Worcester Polytechnic Institute (WPI) The Limits of Geometric Modeling Although graphics cards can render over 10 million polygons per second Many phenomena even more detailed Clouds Grass Terrain Skin Computationally inexpensive way to add details Image complexity does not affect the complexity of geometry processing 2 (transformation, clipping…) Textures in Games Everthing is a texture except foreground characters that require interaction Even details on foreground texture (e.g. clothes) is texture Types of Texturing 1. geometric model 2. texture mapped Paste image (marble) onto polygon Types of Texturing 3. Bump mapping 4. Environment mapping Simulate surface roughness Picture of sky/environment (dimples) over object Texture Mapping 1. Define texture position on geometry 2. projection 4. patch texel 3. texture lookup 3D geometry 2D projection of 3D geometry t 2D image S Texture Representation Bitmap (pixel map) textures: images (jpg, bmp, etc) loaded Procedural textures: E.g. fractal picture generated in .cpp file Textures applied in shaders (1,1) t Bitmap texture: 2D image - 2D array texture[height][width] Each element (or texel ) has coordinate (s, t) s and t normalized to [0,1] range Any (s,t) => [red, green, blue] color s (0,0) Texture Mapping Map? Each (x,y,z) point on object, has corresponding (s, t) point in texture s = s(x,y,z) t = t(x,y,z) (x,y,z) t s texture coordinates world coordinates 6 Main Steps to Apply Texture 1. Create texture object 2. Specify the texture Read or generate image assign to texture (hardware) unit enable texturing (turn on) 3.
    [Show full text]
  • NVIDIA Opengl Extension Specifications
    NVIDIA OpenGL Extension Specifications NVIDIA OpenGL Extension Specifications NVIDIA Corporation Mark J. Kilgard, editor [email protected] May 21, 2001 1 NVIDIA OpenGL Extension Specifications Copyright NVIDIA Corporation, 1999, 2000, 2001. This document is protected by copyright and contains information proprietary to NVIDIA Corporation as designated in the document. Other OpenGL extension specifications can be found at: http://oss.sgi.com/projects/ogl-sample/registry/ 2 NVIDIA OpenGL Extension Specifications Table of Contents Table of NVIDIA OpenGL Extension Support.............................. 4 ARB_imaging........................................................... 6 ARB_multisample....................................................... 7 ARB_multitexture..................................................... 18 ARB_texture_border_clamp............................................. 19 ARB_texture_compression.............................................. 25 ARB_texture_cube_map................................................. 48 ARB_texture_env_add.................................................. 62 ARB_texture_env_combine.............................................. 65 ARB_texture_env_dot3................................................. 73 ARB_transpose_matrix................................................. 76 EXT_abgr............................................................. 81 EXT_bgra............................................................. 84 EXT_blend_color...................................................... 86 EXT_blend_minmax....................................................
    [Show full text]
  • Environment Mapping
    CS 418: Interactive Computer Graphics Environment Mapping Eric Shaffer Some slides adapted from Angel and Shreiner: Interactive Computer Graphics 7E © Addison-Wesley 2015 Environment Mapping How can we render reflections with a rasterization engine? When shading a fragment, usually don’t know other scene geometry Answer: use texture mapping…. Create a texture of the environment Map it onto mirror object surface Any suggestions how generate (u,v)? Types of Environment Maps Sphere Mapping Classic technique… Not supported by WebGL OpenGL supports sphere mapping which requires a circular texture map equivalent to an image taken with a fisheye lens Sphere Mapping Example 5 Sphere Mapping Limitations Visual artifacts are common Sphere mapping is view dependent Acquisition of images non-trivial Need fisheye lens Or render from fisheye lens Cube maps are easier to acquire Or render Acquiring a Sphere Map…. Take a picture of a shiny sphere in a real environment Or render the environment into a texture (see next slide) Why View Dependent? Conceptually a sphere map is generated like ray-tracing Records reflection under orthographic projection From a given view point What is a drawback of this? Cube Map Cube mapping takes a different approach…. Imagine an object is in a box …and you can see the environment through that box 9 Forming a Cube Map Reflection Mapping How Does WebGL Index into Cube Map? •To access the cube map you compute R = 2(N·V)N-V •Then, in your shader vec4 texColor = textureCube(texMap, R); V R •How does WebGL
    [Show full text]
  • Texture Mapping Effects
    CS 543 Computer Graphics Texture Mapping Effects by Cliff Lindsay “Top Ten List” Courtesy of David Letterman’s Late Show and CBS Talk Format List of Texture Mapping Effects from Good to Spectacular (my biased opinion): Highlights: . Define Each Effect . Describe Each Effect Briefly: Theory and Practice. Talk about how each effect extends the idea of general Texture Mapping (previous talk) including Pros and Cons. Demos of selected Texture Mapping Effects 1 Texture Mapping Effect #10 Light Mapping Main idea: Static diffuse lighting contribution for a surface can be captured in a texture and blended with another texture representing surface detail. Highlights: . Eliminate lighting calculation overhead . Light maps are low resolution . Light maps can be applied to multiple textures * = [Images courtesy of flipcode.com] Light Mapping Below is a night scene of a castle. No lighting calculation is being performed at all in the scene. Left: No Light Map applied Right: Light Map applied [Images courtesy of www.gamasutra.com] 2 Texture Mapping Effect #9 Non-Photorealistic Rendering Main idea: Recreating an environment that is focused on depicting a style or communicating a motif as effectively as possible. This is in contrast to Photorealistic which tries to create as real a scene as possible. High lights of NPR: . Toon Shading . Artistic styles (ink, water color, etc.) . Perceptual rendering [Robo Model with and without Toon Shading, Image courtesy of Michael Arias] Non-Photorealistic Rendering Non-Photorealistic Rendering Simple Example: Black
    [Show full text]
  • Polygons Feel No Pain 2014
    Polygons Feel No Pain Volume 1 Ingemar Ragnemalm Course book for TSBK07 Computer Graphics 1 Foreword Why would anyone name a computer graphics book “Polygons feel no pain”? In case you wonder, here are the reasons: • Computer graphics is very much about polygons. • Computer games tend to be violent and I am aiming for gaming! • A major point with it is to inflict less pain (in your back) than the others do. • It is a whole lot more fun than yet another book entitled “Computer Graphics with OpenGL”. So the message is that computer graphics is fun, computer games are fun too, and blasting some grunts in Quake doesn’t hurt anyone. Let’s start the game. INSERT COIN(S) TO CONTINUE... Cover image by Susanne Ragnemalm, based on the original Candide model. Related web pages: http://www.computer-graphics.se http://www.isy.liu.se http://www.ragnemalm.se All content of this book is © Ingemar Ragnemalm 2008-2014. Revised january 2014. 2 1. Introduction This book was written as course material for my course in Computer Graphics at Linköping University. There are many computer graphics books, including the Hearn&Baker book [9] that was course book for our course for many years. That book is excellent in many areas, low-level algorithms and splines in particular. However, the amount of material that was missing for my needs kept growing, and whatever book I con- sidered as replacement seemed to miss just as much. And they are all big books that don’t slip as easily into the backpack as they should.
    [Show full text]
  • CS130 : Computer Graphics Lecture 10: Texture Mapping (Cont.)
    CS130 : Computer Graphics Lecture 10: Texture Mapping (cont.) Tamar Shinar Computer Science & Engineering UC Riverside Perspective correct interpolation Perspective correct interpolation • In assignment 1, we found barycentric coordinates in 2D screen space • but not the correct object space barycentric coords • these coordinates were okay for z-buffer test screen object space space screen object space space Issue: to shade a fragment which is part of a textured triangle we need the barycentric coordinates of the fragment These will be the weights for the weighted average of the vertex texture coordinates. However, after a perspective transformation, the relative distances inside the triangle have been distorted due to foreshortening. I need to get my weights based on object or world space coordinates. Interpolation with screen space weights is incorrect screen object space space correct distorted Perspective correct interpolation Using screen space weights looks wrong for textures [Heckbert and Morton, 1990] http://en.wikipedia.org/wiki/Texture_mapping#Perspective_correctness Do we need to transform back to object space? screen object space space Do we need to transform back to object space? screen object space space NO! <whiteboard> Environment mapping Environment Mapping Use a texture for the distant environment simulate the effect of ray tracing more cheaply Wikimedia Commons Sphere Mapping •Project objects in the environment onto sphere centered at eye •unwrap and store as texture •use reflection direction to lookup texture value How
    [Show full text]