Digital Ornament

Total Page:16

File Type:pdf, Size:1020Kb

Digital Ornament Digital Ornament John Elys University of Wales Institute Cardiff / University of Lincoln Abstract Gaming software has a history of fostering development of economical and creative methods to deal with hardware limitations. Traditionally the visual representation of gaming software has been a poor offspring of high-end visualization. In a twist of irony, this paper proposes that game production software leads the way into a new era of physical digital ornament. The toolbox of the rendering engine evolved rapidly between 1974-1985 and it is still today, 20 years later the main component of all visualization programs. The development of the bump map is of particular interest; its evolution into a physical displacement map provides untold opportunities of the appropriation of the 2D image to a physical 3D object. To expose the creative potential of the displacement map, a wide scope of existing displacement usage has been identified: Top2maya is a scientific appropriation, Caruso St John Architects an architectural precedent and Tord Boonje’s use of 2D digital pattern provides us with an artistic production precedent. Current gaming technologies give us an indication of how the resolution of displacement is set to enter an unprecedented level of geometric detail. As modernity was inspired by the machine age, we should be led by current technological advancement and appropriate its usage. It is about a move away from the simplification of structure and form to one that deals with the real possibilities of expanding the dialogue of surface topology. Digital Ornament is a kinetic process rather than static, its intentions lie in returning the choice of bespoke materials back to the Architect, Designer and Artist. Introduction disappeared from the early modern designer’s lexicon, today advanced Gaming and CAD visualization modeling and CNC technologies will software share a diversity of program allow ornament back onto the center usages, each area has its selective stage. There is considerable potential for methodology on which parts of a program the resurrection of surface (D) texture/ it adopts, a particular modeling tool, a decoration within architecture, art and rendering capability or its animation design. capabilities. Software providers have made This area of research originated from provisions for all areas of production, a student request to create a meshed and this has in turn promoted an eclectic screen for a final year CAD visual. The overlap. Advances in gaming resolution easy option was to create a checker will figure heavily in the forthcoming transparency map to represent the years, leading software provision into mesh of the proposed screen, but there a sophisticated dialogue of increased were issues with that method of virtual topology. Historically, decoration representation. A checker transparency 68 would create the effect of a mesh on a advancements of the virtual world cannot planar object, but it would not be able be appropriated to become physical to carry the shadowing properties of a objects. Today, the texture map is far from modeled mesh object. The question arose being just a mere image-based activity; it that there must be a way of harnessing the has evolved into a powerful tool for the technology of D visual representation, in geometric displacement of topology. This this case the transparency map, to create introduces an exiting era of unparalleled a physical artifact rather than an illusion, detail, which in itself is quite ironic, and there must be a quicker method of considering the checkered history of construction. There always is. virtual representation. The appropriation of the visual D representation into a D model is the Texture Maps aim of the current research, and to explore and highlight the possibilities available. With such a large resource of D Why texture map? In the quest for more maps created for image production, the realistic imagery, one of the most frequent possibilities are endless - always a positive criticisms of early-synthesized raster thing for creativity. The basic toolbox images was the extreme smoothness of the render engine is not an exclusive of surfaces – they showed no texture, CAD Visual element; it appears in many bumps, scratches, dirt or fingerprints. genres of virtual representation, employed Realism demands complexity, or at least differently per incarnation. Gaming and the appearance of complexity. Texture animation are examples of the differing mapping is a relativity efficient means usage but they employ a different working to create the appearance of complexity methodology to that of the CAD visual without the tedium of modeling and Despite the differences between CAD rendering every 3D detail. (Heckbert visual elements animation and gaming, they 1986). do share the same rendering components in the creation of the desired effect. The Texture mapping is in essence research breaks down the complexity of an automated process to replicate a usage into the bare components. From complexity of surface. It is an important its early origins in 976, [this is as early historical precedent from which we can as it gets for the modern day Computer] get the first insight into the imaginative the texture map has become not just an processes developed by the early pioneers image-based activity; it has developed of textural reality. To date, the technique the potential to generate complexity of has remained a visual representation that detail once unimaginable. Is it time to avoids the tedium of physically modeling revisit the role that visualization, gaming the surface of D objects. In the early days and computer animation have in the real of virtual representation, a toolbox was world? formed to enhance the plain surfaces of The world of computer generation the computer D model. This was very is often referred to as the virtual world; much an image-based activity, employed to there is no reason why some of the produce an effect of realism. John Elys Digital Ornament 69 In 986, Paul Heckbert of Pixar Bump Map Evolution surveyed the use of texture mapping, and considering the hardware and software Jim Blinns paper “Simulation of available at that time, it was no surprise Wrinkled Surfaces” published in 978 is that there was so much methodology an iconic paper documenting his research developed in the world of visualization. [invention] into the making of what is Over a short 0-year period the major widely known now as a bump map. It components of today’s rendering software signaled for the first time a progressive environment were formed. Between movement towards the appearance of 97 and 985 texture mapping evolved true complexity and realism in the CAD from being a singular method of mapping environment. Pre 978 visualization and color onto a d object, to a multitude of gaming had lived in a flat raster based user-defined attributes. Figure 1 helps to world. illustrate and define the factions of the Blinn employed a technique that typical rendering palette. It is interesting mapped a grayscale D image to the that this collection of ‘effects’ are based surface of the D object. The grayscale on the original collection of advancements map is employed with a lighting module listed by Paul Heckbert in 985, and are to determine the height of the image very much still in use in today’s rendering based displacement, the white end of and animation programs. the grayscale determined the height of the displacement and the arrangement, • Surface Colour [Ed Catmull 97] and graduation of the source image • Specular reflection [James Blinn file determined the overall effect of 976] displacement of the surface. In essence, • Bump mapping [James Blinn978] the grayscale map is a height map used by • Transparency [Geoffrey Gardner the render engine in displacing the pixels 985] in the image rather than displacing the • Diffuse reflection [Gene Miller geometry of the actual object. 98] • Surface Displacement/ shadows maps [James Kajiya 85] (Heckbert. 1986). Figure 1. From left to right the application of a color map to a sphere. Figure 2. Bump mapping example. ACADIA 2006: Synthetic Landscapes Digital Exchange 70 Digital Making: Superficies: Integral and Applied Displacement mapping bump map on the left is heavily reliant on not only the grayscale map but on the lighting employed to define the image Displacement mapping is a further displacement. Bump mapping works well advancement in Blinn’s technique. It is as a means of imaging topology but the directly related to bump mapping in representation becomes more obvious the that it also uses a height map to model closer you get to it with the camera view. surface perturbations. The difference, A Bump map resides somewhere between however, is that displacement mapping image and surface. actually modifies the surface geometry, The image on the right is computed whereas bump mapping only affects using the displacement map method. The the normal vectors. In other words, displacements are worked out by physically displacement mapping adds geometric displacing the geometry of object to that detail to the mesh (Nuydens 00). of the map. In order to keep an effective workflow the actual displacement is not Polygon displacement can provide fully employed as physical polygon model, us with physical [virtual] surfaces of rather it acts as a geometrical code at the geometric complexity. It is true that the rendering stage. The difference in surface level of processing power, along with topology is self-evident; the displacement the complexity of mesh needed to gain appears physical rather than an effect a detailed object, can be overwhelming, and at close quarters, the sphere retains especially for animation and gaming. accurate topological detail. However, the evolution of the bump map into displacement is the initial stage of a Displacement baking process that automates the modeling of highly detailed objects; a process that does Displacement baking is a term used to away with the tedium of having to model describe the conversion of a displacement every little detail.
Recommended publications
  • AWE Surface 1.2 Documentation
    AWE Surface 1.2 Documentation AWE Surface is a new, robust, highly optimized, physically plausible shader for DAZ Studio and 3Delight employing physically based rendering (PBR) metalness / roughness workflow. Using a primarily uber shader approach, it can be used to render materials such as dielectrics, glass and metal. Features Highlight Physically based BRDF (Oren Nayar for diffuse, Cook Torrance, Ashikhmin Shirley and GGX for specular). Micro facet energy loss compensation for the diffuse and transmission lobe. Transmission with Beer-Lambert based absorption. BRDF based importance sampling. Multiple importance sampling (MIS) with 3delight's path traced area light shaders such as the aweAreaPT shader. Explicit Russian Roulette for next event estimation and path termination. Raytraced subsurface scattering with forward/backward scattering via Henyey Greenstein phase function. Physically based Fresnel for both dielectric and metal materials. Unified index of refraction value for both reflection and transmission with dielectrics. An artist friendly metallic Fresnel based on Ole Gulbrandsen model using reflection color and edge tint to derive complex IOR. Physically based thin film interference (iridescence). Shader based, global illumination. Luminance based, Reinhard tone mapping with exposure, temperature and saturation controls. Toggle switches for each lobe. Diffuse Oren Nayar based translucency with support for bleed through shadows. Can use separate front/back side diffuse color and texture. Two specular/reflection lobes for the base, one specular/reflection lobe for coat. Anisotropic specular and reflection (only with Ashikhmin Shirley and GGX BRDF), with map- controllable direction. Glossy Fresnel with explicit roughness values, one for the base and one for the coat layer. Optimized opacity handling with user controllable thresholds.
    [Show full text]
  • Steep Parallax Mapping
    I3D 2005 Posters Session Steep Parallax Mapping Morgan McGuire* Max McGuire Brown University Iron Lore Entertainment N E E E E True Surface I I I I Bump x P Surface P P P Parallax Mapping Steep Parallax Mapping Bump Mapping Parallax Mapping Steep Parallax Mapping Fig. 2 The viewer perceives an intersection at (P) from Fig. 1 A single polygon with a high-frequency bump map. shading, although the true intersection occurred at (I). Abstract. We propose a new bump mapping scheme that We choose t to be the first ti for which NB [ti]α > 1.0 – i / n can produce parallax, self-occlusion, and self-shadowing for and then shade as with other bump mapping methods: arbitrary bump maps yet is efficient enough for games when float step = 1.0 / n implemented in a pixel shader and uses existing data formats. vec2 dt = E.xy * bumpScale / (n * E.z) Related Work float height = 1.0; Let E and L be unit vectors from the eye and light in a local vec2 t = texCoord.xy; vec4 nb = texture2DLod (NB, t, LOD); tangent space. At a pixel, let 2-vector s be the texture coordinate (i.e. where the eye ray hits the true, flat surface) while (nb.a < height) { and t be the coordinate of the texel where the ray would have height -= step; t += dt; nb = texture2DLod (NB, t, LOD); hit the surface if it were actually displaced by the bump map. } Shading is computed by Phong illumination, using the // ... Shade using N = nb.rgb normal to the scalar bump map surface B at t and the color of the texture map at t.
    [Show full text]
  • Real-Time Rendering Techniques with Hardware Tessellation
    Volume 34 (2015), Number x pp. 0–24 COMPUTER GRAPHICS forum Real-time Rendering Techniques with Hardware Tessellation M. Nießner1 and B. Keinert2 and M. Fisher1 and M. Stamminger2 and C. Loop3 and H. Schäfer2 1Stanford University 2University of Erlangen-Nuremberg 3Microsoft Research Abstract Graphics hardware has been progressively optimized to render more triangles with increasingly flexible shading. For highly detailed geometry, interactive applications restricted themselves to performing transforms on fixed geometry, since they could not incur the cost required to generate and transfer smooth or displaced geometry to the GPU at render time. As a result of recent advances in graphics hardware, in particular the GPU tessellation unit, complex geometry can now be generated on-the-fly within the GPU’s rendering pipeline. This has enabled the generation and displacement of smooth parametric surfaces in real-time applications. However, many well- established approaches in offline rendering are not directly transferable due to the limited tessellation patterns or the parallel execution model of the tessellation stage. In this survey, we provide an overview of recent work and challenges in this topic by summarizing, discussing, and comparing methods for the rendering of smooth and highly-detailed surfaces in real-time. 1. Introduction Hardware tessellation has attained widespread use in computer games for displaying highly-detailed, possibly an- Graphics hardware originated with the goal of efficiently imated, objects. In the animation industry, where displaced rendering geometric surfaces. GPUs achieve high perfor- subdivision surfaces are the typical modeling and rendering mance by using a pipeline where large components are per- primitive, hardware tessellation has also been identified as a formed independently and in parallel.
    [Show full text]
  • Quadtree Displacement Mapping with Height Blending
    Quadtree Displacement Mapping with Height Blending Practical Detailed Multi-Layer Surface Rendering Michal Drobot Technical Art Director Reality Pump Outline • Introduction • Motivation • Existing Solutions • Quad tree Displacement Mapping • Shadowing • Surface Blending • Conclusion Introduction • Next generation rendering – Higher quality per-pixel • More effects • Accurate computation – Less triangles, more sophistication • Ray tracing • Volumetric effects • Post processing – Real world details • Shadows • Lighting • Geometric properties Surface rendering • Surface rendering stuck at – Blinn/Phong • Simple lighting model – Normal mapping • Accounts for light interaction modeling • Doesn’t exhibit geometric surface depth – Industry proven standard • Fast, cheap, but we want more… Terrain surface rendering • Rendering terrain surface is costly – Requires blending • With current techniques prohibitive – Blend surface exhibit high geometric complexity Surface properties • Surface geometric properties – Volume – Depth – Various frequency details • Together they model visual clues – Depth parallax – Self shadowing – Light Reactivity Surface Rendering • Light interactions – Depends on surface microstructure – Many analytic solutions exists • Cook Torrance BDRF • Modeling geometric complexity – Triangle approach • Costly – Vertex transform – Memory • More useful with Tessellation (DX 10.1/11) – Ray tracing Motivation • Render different surfaces – Terrains – Objects – Dynamic Objects • Fluid/Gas simulation • Do it fast – Current Hardware – Consoles
    [Show full text]
  • Texture Mapping
    Texture Mapping Slides from Rosalee Wolfe DePaul University http://www.siggraph.org/education/materials/HyperGraph/mapping/r_wolf e/r_wolfe_mapping_1.htm 1 1. Mapping techniques add realism and interest to computer graphics images. Texture mapping applies a pattern of color to an object. Bump mapping alters the surface of an object so that it appears rough, dented or pitted. In this example, the umbrella, background, beachball and beach blanket have texture maps. The sand has been bump mapped. These and other mapping techniques are the subject of this slide set. 2 2: When creating image detail, it is cheaper to employ mapping techniques that it is to use myriads of tiny polygons. The image on the right portrays a brick wall, a lawn and the sky. In actuality the wall was modeled as a rectangular solid, and the lawn and the sky were created from rectangles. The entire image contains eight polygons.Imagine the number of polygon it would require to model the blades of grass in the lawn! Texture mapping creates the appearance of grass without the cost of rendering thousands of polygons. 3 3: Knowing the difference between world coordinates and object coordinates is important when using mapping techniques. In object coordinates the origin and coordinate axes remain fixed relative to an object no matter how the object’s position and orientation change. Most mapping techniques use object coordinates. Normally, if a teapot’s spout is painted yellow, the spout should remain yellow as the teapot flies and tumbles through space. When using world coordinates, the pattern shifts on the object as the object moves through space.
    [Show full text]
  • Towards a Set of Techniques to Implement Bump Mapping
    Towards a set of techniques to implement Bump Mapping Márcio da Silva Camilo, Bernardo Nogueira S. Hodge, Rodrigo Pereira Martins, Alexandre Sztajnberg Departamento de Informática e Ciências da Computação Universidade Estadual do Rio de Janeiro {pmacstronger, bernardohodge , rodrigomartins , alexszt}@ime.uerj.br Abstract . irregular lighting appearance without modeling its irregular patterns as true geometric perturbations, The revolution of three dimensional video games increasing the model’s polygon count, hence decreasing lead to an intense development of graphical techniques applications’ performance. and hardware. Texture-mapping hardware is now able to There are several other techniques for generate interactive computer-generated imagery with implementing bump mapping. Some of them, such as high levels of per-pixel detail. Nevertheless, traditional Emboss Mapping, do not result in a good appearance for single texture techniques are not able to simulate bumped some surfaces, in general because they use a too simplistic surfaces decently. More traditional bump mapping approximation of light calculation [4]. Others, such as techniques, such as Emboss bump mapping, most times Blinn’s original bump mapping idea, are computationally deploy an undesirable or ‘fake’ appearance to wrinkles in expensive to calculate in real-time [7]. Dot3 bump surfaces. Bump mapping can be applied to different types mapping deploys a great final result on surface appearance of applications varying form computer games, 3d and is feasible in today’s hardware in a single rendering environment simulations, and architectural projects, pass. It is based on a mathematical model of lighting among others. In this paper we will examine a method that intensity and reflection calculated for each pixel on a can be applied in 3d game engines, which uses texture- surface to be rendered.
    [Show full text]
  • Dynamic Image-Space Per-Pixel Displacement Mapping With
    ThoseThose DeliciousDelicious TexelsTexels…… or… DynamicDynamic ImageImage--SpaceSpace PerPer--PixelPixel DisplacementDisplacement MappingMapping withwith SilhouetteSilhouette AntialiasingAntialiasing viavia ParallaxParallax OcclusionOcclusion MappingMapping Natalya Tatarchuk 3D Application Research Group ATI Research, Inc. OverviewOverview ofof thethe TalkTalk > The goal > Overview of approaches to simulate surface detail > Parallax occlusion mapping > Theory overview > Algorithm details > Performance analysis and optimizations > In the demo > Art considerations > Uses and future work > Conclusions Game Developer Conference, San Francisco, CA, March 2005 2 WhatWhat ExactlyExactly IsIs thethe Problem?Problem? > We want to render very detailed surfaces > Don’t want to pay the price of millions of triangles > Vertex transform cost > Memory > Want to render those detailed surfaces correctly > Preserve depth at all angles > Dynamic lighting > Self occlusion resulting in correct shadowing > Thirsty graphics cards want more ALU operations (and they can handle them!) > Of course, this is a balancing game - fill versus vertex transform – you judge for yourself what’s best Game Developer Conference, San Francisco, CA, March 2005 3 ThisThis isis NotNot YourYour TypicalTypical ParallaxParallax Mapping!Mapping! > Parallax Occlusion Mapping is a new technique > Different from > Parallax Mapping > Relief Texture Mapping > Per-pixel ray tracing at its core > There are some very recent similar techniques > Correctly handles complicated viewing phenomena
    [Show full text]
  • Per-Pixel Displacement Mapping with Distance Functions
    108_gems2_ch08_new.qxp 2/2/2005 2:20 PM Page 123 Chapter 8 Per-Pixel Displacement Mapping with Distance Functions William Donnelly University of Waterloo In this chapter, we present distance mapping, a technique for adding small-scale dis- placement mapping to objects in a pixel shader. We treat displacement mapping as a ray-tracing problem, beginning with texture coordinates on the base surface and calcu- lating texture coordinates where the viewing ray intersects the displaced surface. For this purpose, we precompute a three-dimensional distance map, which gives a measure of the distance between points in space and the displaced surface. This distance map gives us all the information necessary to quickly intersect a ray with the surface. Our algorithm significantly increases the perceived geometric complexity of a scene while maintaining real-time performance. 8.1 Introduction Cook (1984) introduced displacement mapping as a method for adding small-scale detail to surfaces. Unlike bump mapping, which affects only the shading of surfaces, displacement mapping adjusts the positions of surface elements. This leads to effects not possible with bump mapping, such as surface features that occlude each other and nonpolygonal silhouettes. Figure 8-1 shows a rendering of a stone surface in which occlusion between features contributes to the illusion of depth. The usual implementation of displacement mapping iteratively tessellates a base sur- face, pushing vertices out along the normal of the base surface, and continuing until 8.1 Introduction 123 Copyright 2005 by NVIDIA Corporation 108_gems2_ch08_new.qxp 2/2/2005 2:20 PM Page 124 Figure 8-1. A Displaced Stone Surface Displacement mapping (top) gives an illusion of depth not possible with bump mapping alone (bottom).
    [Show full text]
  • Analytic Displacement Mapping Using Hardware Tessellation
    Analytic Displacement Mapping using Hardware Tessellation Matthias Nießner University of Erlangen-Nuremberg and Charles Loop Microsoft Research Displacement mapping is ideal for modern GPUs since it enables high- 1. INTRODUCTION frequency geometric surface detail on models with low memory I/O. How- ever, problems such as texture seams, normal re-computation, and under- Displacement mapping has been used as a means of efficiently rep- sampling artifacts have limited its adoption. We provide a comprehensive resenting and animating 3D objects with high frequency surface solution to these problems by introducing a smooth analytic displacement detail. Where texture mapping assigns color to surface points at function. Coefficients are stored in a GPU-friendly tile based texture format, u; v parameter values, displacement mapping assigns vector off- and a multi-resolution mip hierarchy of this function is formed. We propose sets. The advantages of this approach are two-fold. First, only the a novel level-of-detail scheme by computing per vertex adaptive tessellation vertices of a coarse (low frequency) base mesh need to be updated factors and select the appropriate pre-filtered mip levels of the displace- each frame to animate the model. Second, since the only connec- ment function. Our method obviates the need for a pre-computed normal tivity data needed is for the coarse base mesh, significantly less map since normals are directly derived from the displacements. Thus, we space is needed to store the equivalent highly detailed mesh. Fur- are able to perform authoring and rendering simultaneously without typical ther space reductions are realized by storing scalar, rather than vec- displacement map extraction from a dense triangle mesh.
    [Show full text]
  • Per-Pixel Vs. Per-Vertex Real-Time Displacement Mapping Riccardo Loggini
    Per-Pixel vs. Per-Vertex Real-Time Displacement Mapping Riccardo Loggini Introduction Displacement Mapping is a technique that provides geometric details to parametric surfaces. The displacement values are stored in grey scale textures called height maps. We can distinguish two main ways to implement such features which are the Per-Pixel and the Per-Vertex approach. + = Macrostructure Surface Height Map Mesostructure Surface Per-Pixel Approach Per-Vertex Approach Per-Pixel displacement mapping on the GPU can be conceptually Per-Vertex displacement mapping on the GPU increase the inter- seen as an optical illusion: the sense of depth is given by displacing nal mesh geometry by a tessellation process and then uses the the mesh texture coordinates that the viewer is looking at. vertices as control points to displace the surface S along its nor- ~ This approach translates to a ray tracing problem. mal direction nS based on the height map h values. visible surface ~ nS (u0, v 0) ~v h ~S ~u S~ (u, v)=~S (u, v)+h (u, v) n~ (u, v) The execution is made by the fragment shader. Most of the algo- 0 ⇤ S rithms of this family use ray marching to find the intersection point The execution is made before the rasterization process and it can with the height map h (u, v). One of them is Parallax Occlusion involve vertex, geometry and/or tessellation shaders. The OpenGL Mapping that was used to generate the left mesh in Preliminary 4 tessellation stage was used to generate the right mesh in Prelim- Results below. inary Results below with a uniform quad tessellation level.
    [Show full text]
  • INTEGRATING PIXOLOGIC ZBRUSH INTO an AUTODESK MAYA PIPELINE Micah Guy Clemson University, [email protected]
    Clemson University TigerPrints All Theses Theses 8-2011 INTEGRATING PIXOLOGIC ZBRUSH INTO AN AUTODESK MAYA PIPELINE Micah Guy Clemson University, [email protected] Follow this and additional works at: https://tigerprints.clemson.edu/all_theses Part of the Computer Sciences Commons Recommended Citation Guy, Micah, "INTEGRATING PIXOLOGIC ZBRUSH INTO AN AUTODESK MAYA PIPELINE" (2011). All Theses. 1173. https://tigerprints.clemson.edu/all_theses/1173 This Thesis is brought to you for free and open access by the Theses at TigerPrints. It has been accepted for inclusion in All Theses by an authorized administrator of TigerPrints. For more information, please contact [email protected]. INTEGRATING PIXOLOGIC ZBRUSH INTO AN AUTODESK MAYA PIPELINE A Thesis Presented to the Graduate School of Clemson University In Partial Fulfillment of the Requirements for the Degree Master of Fine Arts in Digital Production Arts by Micah Carter Richardson Guy August 2011 Accepted by: Dr. Timothy Davis, Committee Chair Dr. Donald House Michael Vatalaro ABSTRACT The purpose of this thesis is to integrate the use of the relatively new volumetric pixel program, Pixologic ZBrush, into an Autodesk Maya project pipeline. As ZBrush is quickly becoming the industry standard in advanced character design in both film and video game work, the goal is to create a succinct and effective way for Maya users to utilize ZBrush files. Furthermore, this project aims to produce a final film of both valid artistic and academic merit. The resulting work produced a guide that followed a Maya-created film project where ZBrush was utilized in the creation of the character models, as well as noting the most useful formats and resolutions with which to approach each file.
    [Show full text]
  • LEAN Mapping
    LEAN Mapping Marc Olano∗ Dan Bakery Firaxis Games Firaxis Games (a) (b) (c) Figure 1: In-game views of a two-layer LEAN map ocean with sun just off screen to the right, and artist-selected shininess equivalent to a Blinn-Phong specular exponent of 13,777: (a) near, (b) mid, and (c) far. Note the lack of aliasing, even with an extremely high power. Abstract 1 Introduction For over thirty years, bump mapping has been an effective method We introduce Linear Efficient Antialiased Normal (LEAN) Map- for adding apparent detail to a surface [Blinn 1978]. We use the ping, a method for real-time filtering of specular highlights in bump term bump mapping to refer to both the original height texture that and normal maps. The method evaluates bumps as part of a shading defines surface normal perturbation for shading, and the more com- computation in the tangent space of the polygonal surface rather mon and general normal mapping, where the texture holds the ac- than in the tangent space of the individual bumps. By operat- tual surface normal. These methods are extremely common in video ing in a common tangent space, we are able to store information games, where the additional surface detail allows a rich visual ex- on the distribution of bump normals in a linearly-filterable form perience without complex high-polygon models. compatible with standard MIP and anisotropic filtering hardware. The necessary textures can be computed in a preprocess or gener- Unfortunately, bump mapping has serious drawbacks with filtering ated in real-time on the GPU for time-varying normal maps.
    [Show full text]