Image-Based Rendering for Volume Visualization by Using Principle Component Analysis

Total Page:16

File Type:pdf, Size:1020Kb

Image-Based Rendering for Volume Visualization by Using Principle Component Analysis Image-Based Rendering for Volume Visualization by Using Principle Component Analysis by Yinghan Xu, B.Sc. Dissertation Presented to the University of Dublin, Trinity College in fulfillment of the requirements for the Degree of Master of Science in Computer Science University of Dublin, Trinity College September 2017 Declaration I, the undersigned, declare that this work has not previously been submitted as an exercise for a degree at this, or any other University, and that unless otherwise stated, is my own work. Yinghan Xu August 28, 2017 Permission to Lend and/or Copy I, the undersigned, agree that Trinity College Library may lend or copy this thesis upon request. Yinghan Xu August 28, 2017 Acknowledgments I am heartily thankful to my supervisor Dr. John Dingliana for his guidance and support throughout my masters study which set up a strong understanding of the subject, which led to the completion of this thesis. I would like to show my gratitude to all my classmates for helping me throughout the master. I would also like to thank my parents for their support, trust, and encouragement. Last but not the least, I would like to acknowledge and appreciate my girlfriend Xinyue, for spending her time to proofread this dissertation. Yinghan Xu University of Dublin, Trinity College September 2017 iv Image-Based Rendering for Volume Visualization by Using Principle Component Analysis Yinghan Xu, M.Sc. University of Dublin, Trinity College, 2017 Supervisor: John Dingliana This dissertation investigates and analyses imrovement to a PCA approach for image- based rendering in C++ based on paper [2]. This technique allows to visualize volume data with lower GPU requirement and lower computing complicity than direct volume rendering. First, eigenspaces were calculated with a volume data set, pre-rendered using a standard ray-caster, from a spherically distributed range of camera positions. The parallel programming with CUDA was used to speed up the reconstruction process. Then, PCA function in OpenCV is used to calculate eigenspaces and reconstruct results. Artefacts in the result of paper [2] were smoothed by MedianBlur with small computing costs. Results quality, memory cost and frame rate were tested and evaluated under different conditions. The result with different resolution of images, cells dimension, number of eigenvectors were given for real-world application. v Contents Acknowledgments iv Abstract v List of Figures ix Chapter 1 Introduction1 1.1 Motivation....................................2 1.2 Methodology..................................3 1.3 Contribution...................................3 1.4 Summary of Chapters.............................4 Chapter 2 Background5 2.1 Image-based rendering.............................5 2.2 Principle component analysis.........................7 2.3 Volume rendering................................9 2.3.1 Direct volume rendering........................9 2.3.2 Image-based volume rendering..................... 10 2.4 Parallel programming.............................. 12 2.4.1 CUDA.................................. 12 2.4.2 OpenCL................................. 12 vi Chapter 3 Design and Implementation 14 3.1 Basic PCA implementation........................... 14 3.1.1 Put the observed object into the matrix............... 14 3.1.2 Subtract the mean........................... 15 3.1.3 Calculate the covariance matrix.................... 15 3.1.4 Calculate the Eigenvectors and Eigenvalues.............. 16 3.1.5 Sort Eigenvalues............................ 16 3.1.6 Choose the number of components and calculate scores....... 16 3.1.7 Reconstruct the original image..................... 17 3.1.8 Whole image PCA and cell-based image PCA............ 17 3.2 Reconstruction with CPU........................... 18 3.2.1 Crop to cell and put the image pixel value into the matrix..... 18 3.2.2 Calculate eigenvectors......................... 18 3.2.3 Reconstruct the original cell image and put cells together to a whole image.................................. 19 3.3 Reconstruction with CUDA (GPU)...................... 19 3.3.1 Parallel programming design for reconstruction........... 19 3.3.2 CUDA Implementation details..................... 20 3.4 Reconstruction with OpenCV......................... 22 3.4.1 Calculate PCA in OpenCV...................... 22 3.4.2 Reconstruction............................. 24 3.4.3 Speed up reconstruction by OpenCL (GPU)............. 25 3.5 Smooth artefacts between cells......................... 26 3.5.1 Mean blur................................ 27 3.5.2 Gaussian Blur.............................. 27 3.5.3 Median Blur............................... 27 3.5.4 Bilateral Blur.............................. 28 vii Chapter 4 Results and Evaluation 29 4.1 Evaluation methods............................... 29 4.1.1 SSIM................................... 30 4.1.2 HDR-VDP-2............................... 30 4.2 Reconstruction................................. 31 4.2.1 Small training images results...................... 31 4.2.2 Reconstruction speed comparison between CPU and GPU (CUDA) 33 4.2.3 Large training images results...................... 33 4.2.4 Reconstruction speed comparison between CPU and GPU (OpenCL) 36 4.3 Smooth artefacts of edges........................... 37 4.3.1 Blur and Median Blur Comparison.................. 37 4.3.2 Kernel size chosen for Median Blur.................. 39 4.3.3 Comparison between smoothed images and unsmoothed images.. 40 4.4 Image quality, frame rate (reconstruction speed) and memory cost..... 43 4.4.1 Same resolution data set........................ 44 4.4.2 Different resolution data set...................... 47 Chapter 5 Conclusions and Future Work 50 5.1 Conclusions................................... 50 5.2 Future work................................... 52 Appendix A Abbreviations 53 Bibliography 55 viii List of Figures 2.1 Parameterization of the Lumigraph[15].....................6 2.2 Projection process (Reconstruction) [14]....................8 2.3 A volume rendered cadaver head using view-aligned texture mapping and diffuse reflection[5]................................9 3.1 Kernel for blur filter[8].............................. 27 4.1 Head training data with 36 images....................... 31 4.2 Left: reconstructed by 15 eigenvectors. Right: reconstructed by 20 eigen- vectors...................................... 32 4.3 Left: reconstructed by 30 eigenvectors. Right: original image........ 32 4.4 Reconstruction speed comparison between normal CPU and CUDA..... 33 4.5 The SSIM value for these four images are all around 0.967. The top left image used 4 eigenvectors with 5*5 cell dimensions. The top right image used 13 eigenvectors with 10*10 cell dimensions. The bottom left image used 40 eigenvectors with 20*20 cell dimensions. The bottom right image used 90 eigenvectors with 30*30 cell dimensions................ 34 4.6 These four zoomed in images are the same part of area (120*120 pixels) corresponding to figure 8 four images...................... 35 4.7 FPS performance comparison between GPU and only CPU with different cells dimension.................................. 36 4.8 Blur and Median Blur Comparison with Kernel Size 1............ 37 ix 4.9 Blur and Median Blur Comparison with Kernel Size 3............ 38 4.10 Blur and Median Blur Comparison with Kernel Size 5............ 38 4.11 Cells dimension are 10*10. Image quality with the different size of kernel in different number of components....................... 39 4.12 Cells dimension are 20*20. Image quality with the different size of kernel in different number of components....................... 39 4.13 Cells dimension are 30*30. Image quality with the different size of kernel in different number of components....................... 40 4.14 The dataset used here is 720*720 head images. The chart compares qual- ities (SSIM) between smoothed image and unsmoothed image in different cells dimension. Reconstruction with 10*10, 20*20 and 30*30 cells are shown here. The kernel size 3 used in smoothed image............ 41 4.15 The dataset used here is 720*720 head images. The chart compares qual- ities (HDR-VDP-2) between smoothed image and unsmoothed image in different cells dimension. Reconstruction with 10*10, 20*20 and 30*30 cells are shown here. The kernel size 3 used in smoothed image....... 41 4.16 Zoomed in (200*200) comparison between unsmoothed images and smoothed images. The left column is unsmoothed images. The right column is smoothed images, whose kernel size is 3. The first row's images are con- structed by 7 eigenvectors for 10*10 cells dimension. The second row's images are constructed by 25 eigenvectors for 20*20 cells dimension. The third row's images are constructed by 60 eigenvectors for 30*30 cells di- mension...................................... 43 4.17 The relationship between number of eigenvectors used and SSIM value... 44 4.18 The relationship between SSIM value and frame per second (720*720 res- olutions). The corresponding used eigenvectors are labeled upon lines.... 45 x 4.19 The relationship between HDR-VDP-2 quality value and frame per second (720*720 resolutions). The corresponding used eigenvectors are labeled upon lines..................................... 46 4.20 The relationship between SSIM value and frame per second (540*540 res- olutions). The corresponding used eigenvectors are labeled upon lines.... 48 4.21 Three different kinds of resolution data set with the same content. 20*20 cells dimension are used............................. 49 A.1 These three pair of
Recommended publications
  • Stardust: Accessible and Transparent GPU Support for Information Visualization Rendering
    Eurographics Conference on Visualization (EuroVis) 2017 Volume 36 (2017), Number 3 J. Heer, T. Ropinski and J. van Wijk (Guest Editors) Stardust: Accessible and Transparent GPU Support for Information Visualization Rendering Donghao Ren1, Bongshin Lee2, and Tobias Höllerer1 1University of California, Santa Barbara, United States 2Microsoft Research, Redmond, United States Abstract Web-based visualization libraries are in wide use, but performance bottlenecks occur when rendering, and especially animating, a large number of graphical marks. While GPU-based rendering can drastically improve performance, that paradigm has a steep learning curve, usually requiring expertise in the computer graphics pipeline and shader programming. In addition, the recent growth of virtual and augmented reality poses a challenge for supporting multiple display environments beyond regular canvases, such as a Head Mounted Display (HMD) and Cave Automatic Virtual Environment (CAVE). In this paper, we introduce a new web-based visualization library called Stardust, which provides a familiar API while leveraging GPU’s processing power. Stardust also enables developers to create both 2D and 3D visualizations for diverse display environments using a uniform API. To demonstrate Stardust’s expressiveness and portability, we present five example visualizations and a coding playground for four display environments. We also evaluate its performance by comparing it against the standard HTML5 Canvas, D3, and Vega. Categories and Subject Descriptors (according to ACM CCS):
    [Show full text]
  • Volume Rendering
    Volume Rendering 1.1. Introduction Rapid advances in hardware have been transforming revolutionary approaches in computer graphics into reality. One typical example is the raster graphics that took place in the seventies, when hardware innovations enabled the transition from vector graphics to raster graphics. Another example which has a similar potential is currently shaping up in the field of volume graphics. This trend is rooted in the extensive research and development effort in scientific visualization in general and in volume visualization in particular. Visualization is the usage of computer-supported, interactive, visual representations of data to amplify cognition. Scientific visualization is the visualization of physically based data. Volume visualization is a method of extracting meaningful information from volumetric datasets through the use of interactive graphics and imaging, and is concerned with the representation, manipulation, and rendering of volumetric datasets. Its objective is to provide mechanisms for peering inside volumetric datasets and to enhance the visual understanding. Traditional 3D graphics is based on surface representation. Most common form is polygon-based surfaces for which affordable special-purpose rendering hardware have been developed in the recent years. Volume graphics has the potential to greatly advance the field of 3D graphics by offering a comprehensive alternative to conventional surface representation methods. The object of this thesis is to examine the existing methods for volume visualization and to find a way of efficiently rendering scientific data with commercially available hardware, like PC’s, without requiring dedicated systems. 1.2. Volume Rendering Our display screens are composed of a two-dimensional array of pixels each representing a unit area.
    [Show full text]
  • Efficiently Using Graphics Hardware in Volume Rendering Applications
    Efficiently Using Graphics Hardware in Volume Rendering Applications Rudiger¨ Westermann, Thomas Ertl Computer Graphics Group Universitat¨ Erlangen-Nurnber¨ g, Germany Abstract In this paper we are dealing with the efficient generation of a visual representation of the information present in volumetric data OpenGL and its extensions provide access to advanced per-pixel sets. For scalar-valued volume data two standard techniques, the operations available in the rasterization stage and in the frame rendering of iso-surfaces, and the direct volume rendering, have buffer hardware of modern graphics workstations. With these been developed to a high degree of sophistication. However, due to mechanisms, completely new rendering algorithms can be designed the huge number of volume cells which have to be processed and and implemented in a very particular way. In this paper we extend to the variety of different cell types only a few approaches allow the idea of extensively using graphics hardware for the rendering of parameter modifications and navigation at interactive rates for real- volumetric data sets in various ways. First, we introduce the con- istically sized data sets. To overcome these limitations we provide cept of clipping geometries by means of stencil buffer operations, a basis for hardware accelerated interactive visualization of both and we exploit pixel textures for the mapping of volume data to iso-surfaces and direct volume rendering on arbitrary topologies. spherical domains. We show ways to use 3D textures for the ren- Direct volume rendering tries to convey a visual impression of dering of lighted and shaded iso-surfaces in real-time without ex- the complete 3D data set by taking into account the emission and tracting any polygonal representation.
    [Show full text]
  • A Survey of Algorithms for Volume Visualization
    A Survey of Algorithms for Volume Visualization T. Todd Elvins Advanced Scientific Visualization Laboratory San Diego Supercomputer Center "... in 10 years, all rendering will be volume rendering." Jim Kajiya at SIGGRAPH '91 Many computer graphics programmers are working in the area is given in [Fren89]. Advanced topics in medical volume of scientific visualization. One of the most interesting and fast- visualization are covered in [Hohn90][Levo90c]. growing areas in scientific visualization is volume visualization. Volume visualization systems are used to create high-quality Furthering scientific insight images from scalar and vector datasets defined on multi- dimensional grids, usually for the purpose of gaining insight into a This section introduces the reader to the field of volume scientific problem. Most volume visualization techniques are visualization as a subfield of scientific visualization and discusses based on one of about five foundation algorithms. These many of the current research areas in both. algorithms, and the background necessary to understand them, are described here. Pointers to more detailed descriptions, further Challenges in scientific visualization reading, and advanced techniques are also given. Scientific visualization uses computer graphics techniques to help give scientists insight into their data [McCo87] [Brod91]. Introduction Insight is usually achieved by extracting scientifically meaningful information from numerical descriptions of complex phenomena The following is an introduction to the fast-growing field of through the use of interactive imaging systems. Scientists need volume visualization for the computer graphics programmer. these systems not only for their own insight, but also to share their Many computer graphics techniques are used in volume results with their colleagues, the institutions that support the visualization.
    [Show full text]
  • Statistically Quantitative Volume Visualization
    Statistically Quantitative Volume Visualization Joe M. Kniss∗ Robert Van Uitert† Abraham Stephens‡ Guo-Shi Li§ Tolga Tasdizen University of Utah National Institutes of Health University of Utah University of Utah University of Utah Charles Hansen¶ University of Utah Abstract Visualization users are increasingly in need of techniques for assessing quantitative uncertainty and error in the im- ages produced. Statistical segmentation algorithms compute these quantitative results, yet volume rendering tools typi- cally produce only qualitative imagery via transfer function- based classification. This paper presents a visualization technique that allows users to interactively explore the un- certainty, risk, and probabilistic decision of surface bound- aries. Our approach makes it possible to directly visual- A) Transfer Function-based Classification B) Unsupervised Probabilistic Classification ize the combined ”fuzzy” classification results from multi- ple segmentations by combining these data into a unified Figure 1: A comparison of transfer function-based classification ver- probabilistic data space. We represent this unified space, sus data-specific probabilistic classification. Both images are based the combination of scalar volumes from numerous segmen- on T1 MRI scans of a human head and show fuzzy classified white- tations, using a novel graph-based dimensionality reduction matter, gray-matter, and cerebro-spinal fluid. Subfigure A shows the scheme. The scheme both dramatically reduces the dataset results of classification using a carefully designed 2D transfer func- size and is suitable for efficient, high quality, quantitative tion based on data value and gradient magnitude. Subfigure B shows visualization. Lastly, we show that the statistical risk aris- a visualization of the data classified using a fully automatic, atlas- ing from overlapping segmentations is a robust measure for based method that infers class statistics using minimum entropy, visualizing features and assigning optical properties.
    [Show full text]
  • Volume Rendering
    Computer Graphics - Volume Rendering - Hendrik Lensch [a couple of slides thanks to Holger Theisel] Computer Graphics WS07/08 – Volume Rendering Overview • Last Week – Subdivision Surfaces • on Sunday – Ida Helene • Today – Volume Rendering • until tomorrow: Evaluate this lecture on http://frweb.cs.uni-sb.de/03.Studium/08.Eva/ Computer Graphics WS07/08 – Volume Rendering 2 Motivation • Applications – Fog, smoke, clouds, fire, water, … – Scientific/medical visualization: CT, MRI – Simulations: Fluid flow, temperature, weather, ... – Subsurface scattering • Effects in Participating Media – Absorption – Emission –Scattering • Out-scattering • In-scattering • Literature – Klaus Engel et al., Real-time Volume Graphics, AK Peters – Paul Suetens, Fundamentals of Medical Imaging, Cambridge University Press Computer Graphics WS07/08 – Volume Rendering Motivation Volume Rendering • Examples of volume visualization: 4 Computer Graphics WS07/08 – Volume Rendering Direct Volume Rendering Computer Graphics WS07/08 – Volume Rendering Volume Acquisition Computer Graphics WS07/08 – Volume Rendering Direct Volume Rendering 7 Computer Graphics WS07/08 – Volume Rendering Direct Volume Rendering • Shear-Warp factorization (Lacroute/Levoy 94) 8 Computer Graphics WS07/08 – Volume Rendering Volume Representations • Cells and voxels voxels: represent cells: represent homogeneous areas inhomogeneous areas 9 Computer Graphics WS07/08 – Volume Rendering Volume Representations • Cells and voxels voxels: represent cells: represent homogeneous areas inhomogeneous areas
    [Show full text]
  • Scientific Visualization
    Viisualliizatiion Heiightt Fiiellds and Conttourrs Scallarr Fiiellds Vollume Renderriing Vecttorr Fiiellds Tensorr Fiiellds and ottherr hiigh--D datta [[Angell Ch.. 12]] Sciientiifiic Viisualliizatiion • Generally do not start with a 3D model • Must deal with very large data sets – MRI, e.g. 512 e 512 e 200 S 50MB points – Visible Human 512 e 512 e 1734 S 433 MB points · Visualize both real-world and simulation data · User interaction · Automatic search 1 Types of Data · Scalar fields (3D volume of scalars) ± E.g., x-ray densities (MRI, CT scan) · Vector fields (3D volume of vectors) ± E.g., velocities in a wind tunnel · Tensor fields (3D volume of tensors [matrices]) ± E.g., stresses in a mechanical part [Angel 12.7] · Static or through time Examplle: Turbullent convectiion · Penetrative turbulent convection in a compressible ideal gas (PTCC) · http://www.vets.ucar.edu/vg/PTCC/index.shtml plot of enstrophy (vorticity squared) ¡ 2 Examplle: Turbullent convectiion Examplle: Turbullent convectiion How do we visualize volume data like this? 3 Heiight Fiielld · Visualizing an explicit function z = f(x,y) · Adding contour curves g(x,y) = c Meshes · Function is sampled (given) at xi, yi, 0 w i, j w n · Assume equally spaced · Generate quadrilateral or triangular mesh · [Asst 1] 4 Contour Curves · Recall: implicit curve f(x,y) = 0 · f(x,y) < 0 inside, f(x,y) > 0 outside · Here: contour curve at f(x,y) = c · Sample at regular intervals for x,y · How can we draw the curve? Marchiing Squares · Sample function f at every grid point xi, yj · For
    [Show full text]
  • Volume and Surface Rendering of 3D Medical Datasets in Unity®
    Volume and Surface Rendering of 3D Medical Datasets in Unity® Soraia Figueiredo Paulo1, Miguel Belo1, Rafael Kuffner dos Anjos1,2,3, Joaquim Armando Jorge1,2, Daniel Simões Lopes1,2 1 INESC-ID, Lisboa, Portugal 2 Instituto Superior Técnico, Universidade de Lisboa, Lisboa, Portugal 3 FCSH, Universidade Nova de Lisboa REngine consists of a complete Unity (C#) project for interactive volume rendering of 3D Textures and surface rendering of triangular meshes. Volume data is reconstructed from medical image datasets, such as Computed Tomography (CT) or Magnetic Resonance Imaging (MRI), and rendered using a raymarching shader. Surface renderings are created based on 3D meshes and traditional shading algorithms such as Blinn-Phong. This tutorial is a step-by-step demonstration of how to use this project to generate a 3D volume from 2D medical images and render 3D surfaces using Unity 3D game engine (https://unity3d.com/). 1. Download project REngine is available for download at the following public repository https://bitbucket.org/it- medex/volume-rendering-engine. On the left menu, click on Downloads (Figure 1) and then, Download Repository. Figure 1. REngine Bitbucket Repository 2. Import into Unity After downloading the .zip file, uncompress the project and open it using Unity. To do so, open the Assets folder and double-click on scene (Figure 2). 1 Figure 2. REngine Unity Project – Assets folder By default, this project will display the volume rendering of a CT dataset (once freely available at the OsiriX DICOM Image Library (https://www.osirix-viewer.com/resources/dicom-image-library/, alias name PHENIX) that uses the mapping function contained within the TransferFunction.txt file (Figure 3).
    [Show full text]
  • Volume Rendering Is Given
    Fast Visualization by Shear-Warp using Spline Models for Data Reconstruction Inauguraldissertation zur Erlangung des akademischen Grades eines Doktors der Naturwissenschaften der Universit¨at Mannheim vorgelegt von Gregor Schlosser aus Cosel Mannheim, 2009 Dekan: Professor Dr. Felix Freiling, Universit¨at Mannheim Referent: Professor Dr. J¨urgen Hesser, Universit¨at Mannheim Korreferent: Professor Dr. Reinhard M¨anner, Universit¨at Mannheim Tag der m¨undlichen Pr¨ufung: 11.02.2009 Acknowledgements First of all I am very thankful to Prof. Reinhard M¨anner the head of the Chair of Computer Science V at the University of Mannheim and the director of the Institute of Computational Medicine (ICM). I am especially grateful to Prof. J¨urgen Hesser who gave me the chance to work and to write my dissertation at the ICM. He introduced me into the attractive field of computer graphics and proposed to efficiently and accurately visualize three-dimensional medical data sets by using the high performance shear-warp method and trivariate cubic and quadratic spline models, respectively. During my work at the ICM I always found a sympathetic and open ear with him, he gave me advice and support and in innervating discussions he helped me to master my apparently desperate problems. I am also obliged because of his precious comments regarding the initial versions of this thesis. I am also very thankful to Dr. Frank Zeilfelder who introduced me into the difficult field of multivariate spline theory. He made the complex theory understandable from a practical point of view. His accuracy and carefulness made me favorably impressed. In this context I also feel obliged to Dr.
    [Show full text]
  • A Close Review of Scientific Visualization Technique & Volume
    International Journal For Technological Research In Engineering Volume 2, Issue 8, April-2015 ISSN (Online): 2347 - 4718 A CLOSE REVIEW OF SCIENTIFIC VISUALIZATION TECHNIQUE & VOLUME VISUALIZATION TECHNIQUE IN COMPUTER GRAPHICS Narinder Kumar1, Rakesh Kumar2 Dept. of Computer Sci. (Punjabi University Patiala) S.B.S.B. Memorial University College, Sardulgarh ABSTRACT: Computer graphics provide a powerful development in the area of engineering design and and efficient way for visualization communication educational research use three dimensions which has become between different ideas like abstract and concrete ideas, for a central component. The paper contains a basic overview of creating, manipulating, and interacting with these the fundamentals of scientific visualization and volume representations. Beginners don't know techniques or more visualization which is also part of computer graphics with a the visualization – what data we get is most important. To practitioner-oriented review of the latest three dimension arrive at visual result is depend on techniques used in graphics display and visualization techniques. Each topic computer graphics. Visualization is any technique for which is covered in this topic is taken and analyzed are representing diagrams, text, sound, video etc. by written by eminent dabsters in the field. graphically. A integrated collection of computer graphics techniques for visualization's scientific behavior is A. Mapping from computer representations to perceptual presented by Computer Visualization. Visualization is used (visual) in most of field in our daily life and has ever expanding Following figure shows a basic idea of encoding techniques application like medical science, whether forecasting, to enhance human communication and understanding. education, Engineering, interactive multimedia, entertainment, simulation, training etc.
    [Show full text]
  • 14 Acceleration Techniques for Volume Rendering Mark W. Jones
    14.1. INTRODUCTION 253 14 Acceleration Techniques for Volume Rendering Mark W. Jones 14.1 Introduction Volume rendering offers an alternative method for the investigation of three dimensional data to that of surface tiling as described by Jones [1], Lorensen and Cline [2], Wilhelms and Van Gelder [3], Wyvill et. al. [4] and Payne and Toga [5]. Surface tiling can be regarded as giving one particular view of the data set, one which just presents all instances of one value – the threshold value. All other values within the data are ignored and do not contribute to the final image. This is acceptable when the data being visualised contains a surface that is readily understandable, as is the case when viewing objects contained within the data produced by CT scans. In certain circumstances this view alone is not enough to reveal the subtle variations in the data, and for such data sets volume rendering was developed [6, 7, 8, 9]. In this paper the underlying technique employed by volume rendering is given in Section 14.2 presented with the aid of a volume rendering model introduced by Levoy [6]. Section 14.3 examines various other volume rendering models and the differing representations they give of the same data. A more efficient method for sampling volume data is presented in Section 14.4 and acceleration techniques are covered in Section 14.5. In Section 14.6 a thorough comparison is made of many of the more popular acceleration techniques with the more efficient method of Section 14.4. It is shown that the new method offers a 30%-50% improvement in speed over current techniques, with very little image degradation and in most cases no visible degradation.
    [Show full text]
  • Application of 3D Visualization Concept Layer Model for Coal-Bed Methane Index System Du Ping-Pinga, Li Wen-Pinga, Sang Shu-Xuna, Wang Lin-Xiub, Zhou Xiao-Zhia
    Procedia Earth and Planetary Procedia Earth and Planetary Science 1 (2009) 977–981 Science www.elsevier.com/locate/procedia The 6th International Conference on Mining Science & Technology Application of 3D visualization concept layer model for coal-bed methane index system Du Ping-pinga, Li Wen-pinga, Sang Shu-xuna, Wang Lin-xiub, Zhou Xiao-zhia aSchool of Resource and Geoscience China University of Mining & Technology, Xuzhou 221116, China bSchool of Management, China University of Mining & Technology, Xuzhou 221116, China Abstract In the information times, comprehending and monitoring the development of science and technology are one of the most important activities. But the fast increasing in technologic information of the network makes people lose their way who finds that it is difficult to comprehend and master the trend of the technologic development. Visualization technology and its open source tools provide a new vision for people to understand the complex structural and unstructured information. In this paper a method of 3D modeling and visualization of coal bed methane index system is proposed. This method integrates basic coal-bed methane index data, borehole discharging data and interpretation section data of geophysical exploration from a view of multi-source data fusion. The 3D spatial data field model is established by using the spatial interpolation technique and drawing technique of the 3D texture hardware assisted volume rendering direct volume is used to realize the volume visualization, and then the spatial distribution characteristics and inner attribute of strata structures are reflected by true 3D model. The method which uses direct volume rendering techniques in 3D modeling of coal bed methane structure can more accurately reflect the coal gas structures and rapidly establish the data field, thus, it possesses good applied prospects.
    [Show full text]