The Shader Format (.Cfx)

Total Page:16

File Type:pdf, Size:1020Kb

The Shader Format (.Cfx) Shaders allow to mathematically calculate the color of pixels rendered, based on data sent by the engine such as lighting and textures from materials. This means that the full flow from object to shader becomes: Overview The Shader Directory Geometry (.cgf, .cdf, .chr, .skin, .cga) The Shader Format (. Material (.mtl) cfx) Shader Item Extension Files The Common Techniques Keep in mind that any material can have multiple sub-materials - this means that multiple shaders have Shader Editing to be drawn separately via unique draw calls. This is a major reason for keeping the number of materials down, as each additional sub-material requires a separate request to the GPU. Shaders are contained inside the Shaders/ directory, and can be loaded from either Engine/Shaders/ for the default engine shaders, or <asset folder> /Shaders/. A shader is defined by the .cfx extension, and are contained inside Shaders/HWScripts/CryFX/. At engine startup, a search for .cfx files will be performed, resulting in all shader source files being parsed and added to the internal shader registry. The Shader Format (.cfx) CRYENGINE shaders are written in a custom shader format that is very similar to HLSL. Most commonly, materials in the engine will use the Illum shader - as this is already set up to go through the common shader paths. This is defined in Engine/Shaders/HWScripts/CryFX/Illum.cfx. For a more simple example, see the ReferenceImage shader in Engine/Shaders/CryFX/ReferenceImage. cfx. This shader is much more light-weight and is intended to draw a diffuse texture without in-engine lighting affecting or distorting the image. Once a .cfx file has been discovered, it is parsed to detect script information, denoted by the variable name 'Script', typically provided at the top of your shader source file as follows: // Shader global descriptions float Script : STANDARDSGLOBAL < string Script = // Determines if the shader will be made visible to UI elements // For example, if set the shader will be available in the Material Editor "Public;" // If set when SupportsFullDeferredShading is not set, we go through the tiled forward pass // In the tiled forward case, we will execute the pass defined in this file - and not Common*Pass "SupportsDeferredShading;" // Determines that the shader supports fully deferred shading, so will not go through the forward pass "SupportsFullDeferredShading;" >; After the main script declaration, we typically include common headers that we will utilize using the '#include' macro, just like we would in C++. Includes in the engine's shader system are denoted by the .cfi extension. For example: #include "ShadeLib.cfi" ...would automatically find and include the ShadeLib.cfi file inside the Shaders/HWScripts/CryFX/ directory. The common headers are followed by definitions of the techniques and shaders that we will use. For example: vert2FragGeneral MyVertexShader(app2vertGeneral IN) { vert2FragGeneral OUT = (vert2FragGeneral)0; streamPos vertPassPos = (streamPos)0; vs_shared_output(IN, OUT, vertPassPos, true); return OUT; } pixout MyPixelShader(vert2FragGeneral IN) { pixout OUT = (pixout)0; OUT.Color.rgb = GetDiffuseTex(diffuseTex, IN.baseTC); OUT.Color.a = 1; return OUT; } Confidential. © 2009-2015 Crytek GmbH. All Rights Reserved. Page 1 of 3 However, these shaders won't be executed immediately - we have to define the General technique that defines our shaders behavior. This is typically done at the very bottom of the file: technique General < string Script = "TechniqueZ=ZPass;" "TechniqueZPrepass=ZPrepass;" "TechniqueMotionBlur=MotionBlurPass;" "TechniqueCustomRender=CustomRenderPass;" "TechniqueShadowGen=ShadowGen;" "TechniqueDebug=DebugPass;" > { pass p0 { VertexShader = MyVertexShader() GeneralVS; PixelShader = MyPixelShader() GeneralPS; ZEnable = true; ZWriteEnable = true; CullMode = Back; } } In the example above, the Main shader technique is set up to use sub-techniques imported in the engine. This ensures that we go through the common deferred path of the engine, instead of having to rewrite all aspects of each shader every time. The most important aspect for us here, is defining use of the common depth buffer technique, defined in Engine/Shaders/HWScripts/CryFX/CommonZPass.cfi. For more information, see "The Common Techniques" further below. After defining sub-techniques, we defined the main pass of our shader to use in the case of forward rendering. In our example, this means setting up the vertex and pixel shaders we defined above to be called. If you want to go through the forward path at all times, remove the "SupportsDeferredShading;" and "SupportsFullDeferredShading;" strings from the script declaration at the top of the file. Once a shader is done parsing, the engine looks for a .ext file with the same name as our shader (.cfx) inside the Shaders/ directory. This file is optional, and allows for exposing properties that can be applied to each material by the user. For example, the Engine/Shaders/HWScripts/CryFX /Illum.cfx file has an equivalent extension file called Engine/Shaders/Illum.ext. For example, we could add a property like this: Property { Name = %MY_PROPERTY Mask = 0x100 Property (My Property Name) Description (Creates a shader permutation that does fancy stuff!) } Once created, a tick-box appears inside the material editor when a material with your shader is selected. This allows for ticking our property, applying the mask to the material's shader generation flags and letting us catch the use of the property inside one of our shader functions: #if %MY_PROPERTY // Custom property logic here #endif This effectively creates a variation of our shader where a number of options are toggled on or off - usually referred to as a shader permutation. Properties are stored as flags, which is why we need a mask to define the bit that identifies whether or not our property is set. Keep in mind that additional flags beyond properties can be set, the different types are: Flag Description type Engine This type of flag is directly exposed by the engine. Engine flags contain general information like the supported shader model or the platform on which the engine is currently running. Materi The properties that we mentioned above are referred to as material flags, and are evaluated at compile-time, so they cannot be modified al at run-time. These are typically exposed to designers via the Material Editor. Confidential. © 2009-2015 Crytek GmbH. All Rights Reserved. Page 2 of 3 Runtime Runtime flags are dynamically set and unset by the engine during rendering. They make it possible to disable a special feature if it is not required for the current object in the current stage of the rendering pipeline. All available run-time flags are defined in the extension file En gine/Shaders/RunTime.ext. The example we used above referenced common techniques such as ZPass and ZPrePass. These techniques define a common ground between the engine's shaders, ensuring that we re-use functionality and go through similar paths instead of having an inconsistent pipeline. This article will not go into the detail of all of the techniques, but will instead focus on the two that matter most - the depth pass, referred to in the engine as the Z pass. To start, we'll open up Engine/Shaders/HWScripts/CryFX/CommonZPass.cfi. At the very bottom is a definition of the 'ZPass' technique that we ended up referencing above. This defines which shaders to use, in our case we'll focus on: VertexShader = Common_ZPassVS() ZVS; PixelShader = Common_ZPassPS() ZPS; This sets the two functions to use for the vertex and pixel shaders, with the actual functions being defined further up in the file. This knowledge can be used to insert common effects, for example to modify the output color of our material. Note that since the depth pass is included into our source file, we can always reference our material flags to support custom effects that are only activated for our material - instead of affecting all shaders going through the regular pipeline. During development, the engine will automatically compile shader permutations on demand as they are used. However, run-time reloading is also implemented and can be enabled by setting the r_ShadersEditing CVar to a value of 1. This ensures that when a file change is detected in a .cfx file, the shader is automatically recompiled. Additionally, we can set r_ReloadShaders=1 to manually force a reload of all shaders in the scene. Confidential. © 2009-2015 Crytek GmbH. All Rights Reserved. Page 3 of 3.
Recommended publications
  • Evolution of Programmable Models for Graphics Engines (High
    Hello and welcome! Today I want to talk about the evolution of programmable models for graphics engine programming for algorithm developing My name is Natalya Tatarchuk (some folks know me as Natasha) and I am director of global graphics at Unity I recently joined Unity… 4 …right after having helped ship the original Destiny @ Bungie where I was the graphics lead and engineering architect … 5 and lead the graphics team for Destiny 2, shipping this year. Before that, I led the graphics research and demo team @ AMD, helping drive and define graphics API such as DirectX 11 and define GPU hardware features together with the architecture team. Oh, and I developed a bunch of graphics algorithms and demos when I was there too. At Unity, I am helping to define a vision for the future of Graphics and help drive the graphics technology forward. I am lucky because I get to do it with an amazing team of really talented folks working on graphics at Unity! In today’s talk I want to touch on the programming models we use for real-time graphics, and how we could possibly improve things. As all in the room will easily agree, what we currently have as programming models for graphics engineering are rather complex beasts. We have numerous dimensions in that domain: Model graphics programming lives on top of a very fragmented and complex platform and API ecosystem For example, this is snapshot of all the more than 25 platforms that Unity supports today, including PC, consoles, VR, mobile platforms – all with varied hardware, divergent graphics API and feature sets.
    [Show full text]
  • Rendering 13, Deferred Shading, a Unity Tutorial
    Catlike Coding Unity C# Tutorials Rendering 13 Deferred Shading Explore deferred shading. Fill Geometry Bufers. Support both HDR and LDR. Work with Deferred Reflections. This is part 13 of a tutorial series about rendering. The previous installment covered semitransparent shadows. Now we'll look at deferred shading. This tutorial was made with Unity 5.5.0f3. The anatomy of geometry. 1 Another Rendering Path Up to this point we've always used Unity's forward rendering path. But that's not the only rendering method that Unity supports. There's also the deferred path. And there are also the legacy vertex lit and the legacy deferred paths, but we won't cover those. So there is a deferred rendering path, but why would we bother with it? After all, we can render everything we want using the forward path. To answer that question, let's investigate their diferences. 1.1 Switching Paths Which rendering path is used is defined by the project-wide graphics settings. You can get there via Edit / Project Settings / Graphics. The rendering path and a few other settings are configured in three tiers. These tiers correspond to diferent categories of GPUs. The better the GPU, the higher a tier Unity uses. You can select which tier the editor uses via the Editor / Graphics Emulation submenu. Graphics settings, per tier. To change the rendering path, disable Use Defaults for the desired tier, then select either Forward or Deferred as the Rendering Path. 1.2 Comparing Draw Calls I'll use the Shadows Scene from the Rendering 7, Shadows tutorial to compare both approaches.
    [Show full text]
  • Real-Time Lighting Effects Using Deferred Shading
    Real-time Lighting Effects using Deferred Shading Michal Ferko∗ Supervised by: Michal Valient† Faculty of Mathematics, Physics and Informatics Comenius University Bratislava / Slovakia Abstract We are targeting OpenGL 3 capable hardware, because we require the framebuffer object features as well as mul- Rendering realistic objects at interactive frame rates is a tiple render targets. necessary goal for many of today’s applications, especially computer games. However, most rendering engines used in these games induce certain limitations regarding mov- 2 Related Work ing of objects or the amount of lights used. We present a rendering system that helps overcome these limitations There are many implementations of Deferred Shading and while the system is still able to render complex scenes at this concept has been widely used in modern games [15] 60 FPS. Our system uses Deferred Shading with Shadow [12] [5], coupled with techniques used in our paper as well Mapping for a more efficient way to synthesize lighting as certain other. coupled with Screen-Space Ambient Occlusion to fine- Deferred Shading does not directly allow rendering of tune the final shading. We also provide a way to render transparent objects and therefore, we need to use a differ- transparent objects efficiently without encumbering the ent method to render transparent objects. There are several CPU. approaches to hardware-accelerated rendering of transpar- ent objects without the need to sort geometry. This group Keywords: Real-time Rendering, Deferred Shading, of algorithms is referred to as Order-Independent Trans- High-dynamic range rendering, Tone-mapping, Order- parency. Independent Transparency, Ambient Occlusion, Screen- An older approach is Depth Peeling [7] [4], which re- Space Ambient Occlusion, Stencil Routed A-Buffer quires N scene rendering passes to capture N layers of transparent geometry.
    [Show full text]
  • More Efficient Virtual Shadow Maps for Many Lights
    1 More Efficient Virtual Shadow Maps for Many Lights Ola Olsson1;2, Markus Billeter1;3, Erik Sintorn1, Viktor Kampe¨ 1, and Ulf Assarsson1 (Invited Paper) Abstract—Recently, several algorithms have been introduced for shading is much more important. To this end we use that enable real-time performance for many lights in applications Clustered Deferred Shading [3], as our starting point. This such as games. In this paper, we explore the use of hardware- algorithm offers the highest light-culling efficiency among supported virtual cube-map shadows to efficiently implement high-quality shadows from hundreds of light sources in real time current real-time many-light algorithms and the most robust and within a bounded memory footprint. In addition, we explore shading performance. Moreover, clustered shading provides the utility of ray tracing for shadows from many lights and tight 3D bounds around groups of samples in the frame present a hybrid algorithm combining ray tracing with cube buffer and therefore can be viewed as a fast voxelization of maps to exploit their respective strengths. Our solution supports the visible geometry. Thus, as we will show, these clusters real-time performance with hundreds of lights in fully dynamic high-detail scenes. provide opportunities for efficient culling of shadow casters and allocation of shadow map memory. Index Terms—Computer graphics, GPU, real-time shading, shadows, virtual texturing. A. Contributions We contribute an efficient culling scheme, based on clusters, I. INTRODUCTION which is used to render shadow-casting geometry to many cube In recent years, several techniques have been presented shadow maps. We demonstrate that this can enable real-time that enable real-time performance for applications such as rendering performance using shadow maps for hundreds of games using hundreds to many thousands of lights.
    [Show full text]
  • A Novel Multithreaded Rendering System Based on a Deferred Approach
    VIII Brazilian Symposium on Games and Digital Entertainment Rio de Janeiro, RJ – Brazil, October, 8th-10th 2009 A Novel Multithreaded Rendering System based on a Deferred Approach Jorge Alejandro Lorenzon Esteban Walter Gonzalez Clua Universidad Austral, Argentina Media Lab – UFF, Brazil [email protected] [email protected] Figure 1: Mix of the final illuminated picture, the diffuse color buffer and the normal buffer Abstract Therefore, the architecture of newer game engines must include fine-grained multithreaded algorithms This paper presents the architecture of a rendering and systems. Fortunately for some systems like physics system designed for multithreaded rendering. The and AI this can be done. However, when it comes to implementation of the architecture following a deferred rendering there is one big issue: All draw and state rendering approach shows gains of 65% on a dual core calls must go to the graphics processing unit (GPU) in 1 machine. a serialized manner . This limits game engines as only one thread can actually execute draw calls to the Keywords : multithreaded rendering, deferred graphics card. Adding to the problem, draw calls and rendering, DirectX 11, command buffer, thread pool state management of the graphics pipeline are expensive for the CPU as there is a considerable overhead created by the API and driver. For this 1. Introduction reason, most games and 3D applications are CPU Game engines and 3D software are constantly bound and rely on batching 3D models to feed the changing as the underlying hardware and low level GPU. APIs evolve. The main driving force of change is the pursuit of greater performance for 3D software, which Microsoft, aware of this problem, is pushing means, pushing more polygons with more realistic forward a new multithreaded graphics API for the PC, models of illumination and shading techniques to the Direct3D11.
    [Show full text]
  • Efficient Virtual Shadow Maps for Many Lights
    This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TVCG.2015.2418772, IEEE Transactions on Visualization and Computer Graphics 1 More Efficient Virtual Shadow Maps for Many Lights Ola Olsson1;2, Markus Billeter1;3, Erik Sintorn1, Viktor Kampe¨ 1, and Ulf Assarsson1 (Invited Paper) Abstract—Recently, several algorithms have been introduced algorithm offers the highest light-culling efficiency among that enable real-time performance for many lights in applications current real-time many-light algorithms and the most robust such as games. In this paper, we explore the use of hardware- shading performance. Moreover, clustered shading provides supported virtual cube-map shadows to efficiently implement high-quality shadows from hundreds of light sources in real time tight 3D bounds around groups of samples in the frame and within a bounded memory footprint. In addition, we explore buffer and therefore can be viewed as a fast voxelization of the utility of ray tracing for shadows from many lights and the visible geometry. Thus, as we will show, these clusters present a hybrid algorithm combining ray tracing with cube provide opportunities for efficient culling of shadow casters maps to exploit their respective strengths. Our solution supports and allocation of shadow map memory. real-time performance with hundreds of lights in fully dynamic high-detail scenes. A. Contributions I. INTRODUCTION We contribute an efficient culling scheme, based on clusters, In recent years, several techniques have been presented which is used to render shadow-casting geometry to many cube that enable real-time performance for applications such as shadow maps.
    [Show full text]
  • Wipein - F- - a 3D Action Game
    .sssss WipeIn - F- - A 3D Action Game Bachelor's Thesis Computer Science and Engineering Programme CHRISTOPHER ANDERSSON JESPER LINDH MIKAEL MÖLLER MIKAEL OLAISSON KARL SCHMIDT CARL-JOHAN SÖDERSTEN ALLAN WANG Department of Computer Science and Engineering CHALMERS UNIVERSITY OF TECHNOLOGY Gothenburg, Sweden 2011 Bachelor's Thesis DATX11-09 - Rally Sport Racing Game 1 Abstract The following thesis describes a case study of 3D game program- ming. It involves the evaluation of several techniques commonly used in real-time rendering, as well as some associated elds such as mod- elling, collision handling and sound. We will investigate which of the many options available are the most ecient, as well as which areas are preferably put aside, in the aim of achieving an entertaining and visually appealing 3D computer game within a short time span. 2 CONTENTS CONTENTS Contents 1 Introduction 5 1.1 Background . .5 1.2 Purpose . .5 1.3 Problem . .5 1.4 Limitations . .5 1.4.1 Contents . .6 1.4.2 Areas of focus . .6 1.4.3 Open-source code . .6 1.4.4 Computer power . .6 1.5 Method . .6 1.5.1 Choice of programming language and framework . .7 1.5.2 API . .7 1.5.3 Development process . .8 1.6 Game design . .8 2 Graphics 9 2.1 Pipeline . .9 2.2 The application stage . 10 2.3 The geometry stage . 10 2.4 The rasteriser stage . 12 2.4.1 Hidden surface determination . 12 2.5 Shading . 14 2.5.1 The Phong Shading Model . 14 2.5.2 Bidirectional Reectance Distribution Functions .
    [Show full text]
  • Real Shading in Unreal Engine 4 by Brian Karis, Epic Games
    Real Shading in Unreal Engine 4 by Brian Karis, Epic Games Figure 1: UE4: Infiltrator demo Introduction About a year ago, we decided to invest some time in improving our shading model and embrace a more physically based material workflow. This was driven partly by a desire to render more realistic images, but we were also interested in what we could achieve through a more physically based approach to material creation and the use of material layering. The artists felt that this would be an enormous improvement to workflow and quality, and I had already seen these benefits first hand at another studio, where we had transitioned to material layers that were composited offline. One of our technical artists here at Epic experimented with doing the layering in the shader with promising enough results that this became an additional requirement. In order to support this direction, we knew that material layering needed to be simple and effi- cient. With perfect timing came Disney’s presentation [2] concerning their physically based shading and material model used for Wreck-It Ralph. Brent Burley demonstrated that a very small set of material parameters could be sophisticated enough for offline feature film rendering. He also showed that a fairly practical shading model could closely fit most sampled materials. Their work became an inspiration and basis for ours, and like their “principles,” we decided to define goals for our own system: Real-Time Performance • First and foremost, it needs to be efficient to use with many lights visible at a time. 1 Reduced Complexity • There should be as few parameters as possible.
    [Show full text]
  • Deferred Rendering Using Compute Shaders
    Deferred rendering using Compute shaders A comparative study using shader model 4.0 and 5.0 Benjamin Golba 1 | P a g e This thesis is submitted to the Department of Interaction and System Design at Blekinge Institute of Technology in partial fulfillment of the requirements for the Bachelor degree in Computer Science. The thesis is equivalent to 10 weeks of full time studies. Contact Information: Author: Benjamin Golba Address: Folkparksvägen 10:17, 372 40 Ronneby E-mail: [email protected] University advisor: Stefan Petersson Department of Software Engineering and Computer Science Address: Soft Center, RONNEBY Phone: +46 457 38 58 15 Department of Interaction and System Design Blekinge Institute of Technology SE - 372 25 RONNEBY Sweden Internet: http://www.bth.se/tek/ais Phone: +46 457 38 58 00 Fax: +46 457 271 25 2 | P a g e Abstract Game developers today are putting a lot of effort into their games. Consumers are hard to please and demand a game which can provide both fun and visual quality. This is why developers aim to make the most use of what hardware resources are available to them to achieve the best possible quality of the game. It is easy to use too many performance demanding techniques in a game, making the game unplayable. The hard part is to make the game look good without decreasing the performance. This can be done by using techniques in a smart way to make the graphics as smooth and efficient as they can be without compromising the visual quality. One of these techniques is deferred rendering.
    [Show full text]
  • Real-Time 2D Manipulation of Plausible 3D Appearance Using Shading and Geometry Buffers Carlos Jorge Zubiaga Pena
    Real-time 2D manipulation of plausible 3D appearance using shading and geometry buffers Carlos Jorge Zubiaga Pena To cite this version: Carlos Jorge Zubiaga Pena. Real-time 2D manipulation of plausible 3D appearance using shading and geometry buffers. Other [cs.OH]. Université de Bordeaux, 2016. English. NNT : 2016BORD0178. tel-01486698 HAL Id: tel-01486698 https://tel.archives-ouvertes.fr/tel-01486698 Submitted on 10 Mar 2017 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. THESIS PRESENTED AT UNIVERSITE´ DE BORDEAUX ECOLE´ DOCTORALE DE MATHMATIQUES ET D’INFORMATIQUE par Carlos Jorge Zubiaga Pe˜na POUR OBTENIR LE GRADE DE DOCTEUR SPECIALIT´ E´ : INFORMATIQUE Real-time 2D manipulation of plausible 3D appearance using shading and geometry buffers Date de soutenance : 7 November 2016 Devant la commission d’examen compose de : Diego Gutierrez . Professeur, Universidad de Zaragoza . Rapporteur Daniel Sykora´ . Professeur associ´e, Czech Technical University in Prague Rapporteur Pascal Guitton ...... Professeur, Univerist´eBordeaux ................... ..... Pr´esident David Vanderhaeghe Maˆıtre de Conferences, Univerist´ede Toulouse . Examinateur Xavier Granier ...... Professeur, Institut d’Optique .................. Examinateur Pascal Barla ......... Charg´ede recherche, Inria ................... ........... Directeur 2016 Abstract Traditional artists paint directly on a canvas and create plausible appearances of real-world scenes.
    [Show full text]
  • Deferred Shading Tutorial
    Deferred Shading Tutorial Fabio Policarpo1 Francisco Fonseca2 [email protected] [email protected] CheckMate Games1,2 Pontifical Catholic University of Rio de Janeiro2 ICAD/Igames/VisionLab 1. Introduction Techniques usually consider non-interactive a few years ago are now possible in real-time using the flexibility and speed of new programmable graphics hardware. An example of that is the deferred shading technique, which is an approach that postpones shading calculations for a fragment1 until the visibility of that fragment is completely determined. In other words, it implies that only fragments that really contribute to the resultant image are shaded. Although deferred shading has become practical for real-time applications in recent years, this technique was firstly published in 1988 by Michael Deering et al. [Deering88]. In that work, the authors proposed a VLSI system where a pipeline of triangle processors rasterizes the geometry, and then a pipeline of shading processors applies Phong shading [Phong75] with multiple light sources to such geometry. After the initial research performed by Deering et al., the next relevant work involving deferred shading was developed by Saito and Takahashi [Saito90] in 1990. The authors of this article proposed a rendering technique that produces 3D images that favor the recognition of shapes and patterns, since shapes can be readily understood if certain geometric properties are enhanced. In order to optimize the enhancement process, geometric properties of the surfaces are preserved as Geometric-Buffers (G-buffers). So, by using G-buffers as intermediate results, artificial enhancement processes are separated from geometric processes (projection and hidden surface removal) and physical processes (shading and texture mapping), and performed as a post-processing pass.
    [Show full text]
  • Optimization to Deferred Shading Pipeline
    UNIVERSITY OF DUBLIN,TRINITY COLLEGE MASTERS THESIS Optimization to Deferred Shading Pipeline Author: Supervisor: Samir KUMAR Dr. Michael MANZKE August 2017 A thesis submitted in fulfillment of the requirements for the degree of MSc Computer Science (Interactive Entertainment Technology) in the IET School of Computer Science and Statistics iii Declaration of Authorship I, Samir KUMAR, declare that this thesis titled, “Optimization to Deferred Shading Pipeline” and the work presented in it are my own. I confirm that: • This work was done wholly or mainly while in candidature for a master degree at this University. • I declare that this thesis has not been submitted at any other university. • Where I have consulted the published work of others, this is always clearly attributed. • Where I have quoted from the work of others, the source is always given. With the exception of such quotations, this thesis is entirely my own work. • I have acknowledged all main sources of help. • Where the thesis is based on work done by myself jointly with others, I have made clear exactly what was done by others and what I have contributed myself. Signed: Date: v University of Dublin, Trinity College Abstract The Faculty of Engineering, Mathematics and Science School of Computer Science and Statistics MSc Computer Science (Interactive Entertainment Technology) Optimization to Deferred Shading Pipeline by Samir KUMAR Rendering pipelines in the field of graphics are continuously evolved. De- ferred rendering pipeline is a step forward from most Common pipeline Forward rendering pipeline. Deferred rendering pipeline has its disadvan- tages such as its does not support blending.
    [Show full text]