Short167.Pdf (571.2Kb)

Total Page:16

File Type:pdf, Size:1020Kb

Short167.Pdf (571.2Kb) EUROGRAPHICS 2003 / M. Chover, H. Hagen and D. Tost Short Presentations A Framework for Video-based and Hardware-Accelerated Remote 3D-Visualization Frank Goetz, Gitta Domik Computer Graphics, Visualization and Image Processing Fuerstenallee 11, D-33102 Paderborn, Germany (frank.goetz|domik)@uni-paderborn.de Abstract This paper presents a framework for video-based and hardware-accelerated remote 3d-visualization that produces and delivers high quality video streams at an accurate frame rate. The framework is based on the portable scenegraph system OpenSG, the MPEG4IP package and the Apple Darwin Streaming Server. In realtime generated visualizations will be multicast as an ISO MPEG-4 compliant video stream over the RealTime Streaming Protocol from a server to a client computer. On the client computer a Java program and an ISO MPEG-4 compliant video player are used to interact with the delivered visualization. While us- ing MPEG-4 video streams we can achieve high image quality at a low bandwidth. Categories and Subject Descriptors (according to ACM CCS): I.3.1 [Computer Graphics] Distributed/network, C.2.4 [Distributed Systems]: Distributed applications. 1. Introduction A common method to transport visualizations is based Not so long ago the lack of communication was ham- on the streaming of compressed meshes and textures pering the progress of research. Today, researchers from a server to a client computer. While visualizing share information over the internet and cooperation is complex scenes with a huge amount of polygons this demanded in all areas of the industry. Particularly com- method often needs too much network capacity and has puter-generated visualizations took a firm place in many high demands on the capabilities of the client com- fields, like geosciences, medicine, architecture, automo- puters. In most cases, the client computer offers only a bile construction and space technology. fragment of processor power and graphics power that a server is able to achieve. Finally, with this method it is not possible to get accurate frame rates while sending Interactivity and dynamics in realtime generated visu- interactive and realtime generated visualizations from a alizations were strongly improved by today’s more effi- server to a client computer. Accurate frame rate means cient graphics accelerator boards and arising multiproc- at least a frame rate of 20 frames per second. In that essor systems. Often complex and interactive visualiza- case a user has to accept limitations because compro- tions can only be rendered on powerful graphics servers mises were made by sending large datasets to the client or graphics clusters, since only they achieve the desired computer. representation speed and image quality. Therefore it is necessary to transfer the power of these graphics servers to the location and the computer of the user that wants The aim of our work is to develop a platform independ- to work with this visualization. In the last years scien- ent visualization server that takes advantage of the func- tists and researchers have given a great deal of attention tionality of current graphics accelerator boards and de- to the area of remote visualization. livers a high performance video stream to the remote or client computer. In our case high performance video stream means that the picture quality is high, the frame rate is accurate, the needed bandwidth is low and the © The Eurographics Association 2003. Goetz et al / A Framework for Video-based and Hardware-Accelerated Remote 3D-Visualization latency is as short as possible. While using video The generic solution for hardware-accelerated remote streams we have a constant and calculable size of net- visualization from Stegmaier et al.3 works transparently work capacity and processor power that has to be sup- for all OpenGL-based applications and OpenGL-based ported by the client computer. The size of the video scene graphs and does not require any modifications of stream is independent from the size of the visualization existing applications. They use a similar approach as and the graphics accelerator board of the client com- Richardson et al.4 in their paper virtual network comput- puter does not need special features to display interac- ing. In this paper a remote display system that does not tive and animated visualizations at an accurate frame support the remote use of 3d graphics acceleration hard- rate. It is possible to use the majority of available com- ware is presented. puters, even laptops and handhelds, as a visualization client. Ma and Camp5 developed a solution for remote visuali- zation of time-varying data over wide area networks. A further point should be that different users are able to This system involves a display daemon and a display access the data that was generated by the visualization interface. The data from the renderer is automatically server at the same time. Also the coordination between compressed, transported and decompressed. By using a active users should be as easy as possible. It is very custom transport method, they are able to employ arbi- important that every user of the cooperative working trary compression techniques. process gets the same potential for work. There should not be any problems concerning the geographical loca- Other solutions that support remote visualization were tion or the capabilities of the hardware. The user should developed by Engel and Ertl6 and Engel et al.7,8. In tex- not be restricted or excluded from the working process ture-based volume visualization for multiple users on because of using limited hardware or software. The the world wide web Engel and Ertl describe a volume ability to interact, navigate and manipulate at an accu- visualization tool that uses JAVA and the Virtual Real- rate speed has to be available for all users of a coopera- ity Modelling Language (VRML). In this case the client tive working process. computer has to be equipped with a 3d graphics accel- eration hardware to render the transmitted VRML scene The rest of the paper is organized as follows: first, other with an accurate speed. In their latest papers Engel et al. frameworks and solutions for remote 3D-visualization describe visualization systems that use image compres- are introduced; then all necessary packages, libraries sion technologies to transport the visualization data and servers that we need for our solution are presented; from the server to the client computer. In their paper a afterwards, the architecture of our system is explained; framework for interactive hardware-accelerated remote in the last part results are presented. 3d-visualization the visualization parameters and GUI events from the clients are applied to the server applica- tion by sending CORBA (Common Object Request 2. Previous Work Broker Architecture) requests. There are different ways of realizing remote visualiza- tion. One of the latest remote visualization frameworks Their remote visualization framework is very similar to 1 is Chromium . Chromium is a stream-processing our framework. Instead of CORBA we developed our framework for interactive rendering on clusters. own network communication layer. The main difference to all other remote visualization frameworks and also Silicon Graphics, Inc. provides a commercial solution the demarcation to the systems by Engel et al. is that our 2 called OpenGL Vizserver . OpenGL Vizserver is a framework transports MPEG-4 video streams instead of technical and creative computing solution designed to sending only pictures from the server to the client. This deliver visualization and collaboration functionality to improves the quality and capacity of the data to be sent. any client, whether on a desktop workstation or a wire- less tablet. OpenGL Vizserver allows users to remotely view and interact with large data sets from any other 3. Background for our solution system at any location in an organization and to collabo- One requirement of our framework is to use only free rate with multiple colleagues using these same applica- and non commercial packages, libraries and servers. In tions and data. Because of design decisions the OpenGL this chapter all used packages, libraries and servers are Vizserver works only with the SGI Onyx family. Simi- presented briefly. larly to the solution of Stegmaier et al.3, the VizServer relies on dynamically linked executables in order to be able to implant its functionality without modifying the target application. © The Eurographics Association 2003. Goetz et al / A Framework for Video-based and Hardware-Accelerated Remote 3D-Visualization 3.1. CommonCPP program supplying the data. The second device acts as a CommonCPP offers a highly portable C++ application normal video4linux device, it should be usable by any development framework. It provides classes for threads, application that fulfills the video4linux specifications. sockets, daemon management, system logging, object synchronization, realtime network development, persis- The loopback device has two operating modes: tent object management, and file access. These are all features that are not supported by standard C++ librar- In the simple one-copy mode the supplying program ies. specifies the size of the images and the used palette and uses the write function to push its images to the pipe. The CommonCPP Framework is used by our Comm- This mode is mostly for feeding fixed size images with- Server for the communication between the server and out any knowledge about the client. At the moment our the client. The CommServer is a socket based and system only supports this mode and streams a 24bit true thread per session server. For every session a new colour video with a size of 352*288 pixels to every thread will be generated. All incoming messages will be client computer. processed by this thread and after the connection is closed the thread will be terminated. The CommServer In the zero-copy mode the supplying program regularly listens on a standard port and generates a NetRPCMgr polls the device.
Recommended publications
  • Stardust: Accessible and Transparent GPU Support for Information Visualization Rendering
    Eurographics Conference on Visualization (EuroVis) 2017 Volume 36 (2017), Number 3 J. Heer, T. Ropinski and J. van Wijk (Guest Editors) Stardust: Accessible and Transparent GPU Support for Information Visualization Rendering Donghao Ren1, Bongshin Lee2, and Tobias Höllerer1 1University of California, Santa Barbara, United States 2Microsoft Research, Redmond, United States Abstract Web-based visualization libraries are in wide use, but performance bottlenecks occur when rendering, and especially animating, a large number of graphical marks. While GPU-based rendering can drastically improve performance, that paradigm has a steep learning curve, usually requiring expertise in the computer graphics pipeline and shader programming. In addition, the recent growth of virtual and augmented reality poses a challenge for supporting multiple display environments beyond regular canvases, such as a Head Mounted Display (HMD) and Cave Automatic Virtual Environment (CAVE). In this paper, we introduce a new web-based visualization library called Stardust, which provides a familiar API while leveraging GPU’s processing power. Stardust also enables developers to create both 2D and 3D visualizations for diverse display environments using a uniform API. To demonstrate Stardust’s expressiveness and portability, we present five example visualizations and a coding playground for four display environments. We also evaluate its performance by comparing it against the standard HTML5 Canvas, D3, and Vega. Categories and Subject Descriptors (according to ACM CCS):
    [Show full text]
  • Data Management for Augmented Reality Applications
    Technische Universitat¨ Munchen¨ c c c c Fakultat¨ fur¨ Informatik c c c c c c c c c c c c Diplomarbeit Data Management for Augmented Reality Applications ARCHIE: Augmented Reality Collaborative Home Improvement Environment Marcus Tonnis¨ Technische Universitat¨ Munchen¨ c c c c Fakultat¨ fur¨ Informatik c c c c c c c c c c c c Diplomarbeit Data Management for Augmented Reality Applications ARCHIE: Augmented Reality Collaborative Home Improvement Environment Marcus Tonnis¨ Aufgabenstellerin: Prof. Gudrun Klinker, Ph.D. Betreuer: Dipl.-Inf. Martin Bauer Abgabedatum: 15. Juli 2003 Ich versichere, daß ich diese Diplomarbeit selbstandig¨ verfaßt und nur die angegebenen Quellen und Hilfsmittel verwendet habe. Munchen,¨ den 15. Juli 2003 Marcus Tonnis¨ Zusammenfassung Erweiterte Realitat¨ (Augmented Reality, AR) ist eine neue Technologie, die versucht, reale und virtuelle Umgebungen zu kombinieren. Gemeinhin werden Brillen mit eingebauten Computerdisplays benutzt um eine visuelle Er- weiterung der Umgebung des Benutzers zu erreichen. In das Display der Brillen, die Head Mounted Displays genannt werden, konnen¨ virtuelle Objekte projiziert werden, die fur¨ den Benutzer ortsfest erscheinen. Das am Lehrstuhl fur¨ Angewandte Softwaretechnik der Technischen Universitat¨ Munchen¨ angesiedelte Projekt DWARF versucht, Methoden des Software Engineering zu benutzen, um durch wiederverwendbare Komponenten die prototypische Implementierung neuer Komponenten zu beschleunigen. DWARF besteht aus einer Sammlung von Softwarediensten, die auf mobiler verteilter Hardware agieren und uber¨ drahtlose oder fest verbundene Netzwerke miteinander kom- munizieren konnen.¨ Diese Kommunikation erlaubt, personalisierte Komponenten mit einge- betteten Diensten am Korper¨ mit sich zu fuhren,¨ wahrend¨ Dienste des Umfeldes intelligente Umgebungen bereitstellen konnen.¨ Die Dienste erkennen sich gegenseitig und konnen¨ dy- namisch kooperieren, um gewunschte¨ Funktionalitat¨ zu erreichen, die fur¨ Augmented Rea- lity Anwendungen gewunscht¨ ist.
    [Show full text]
  • Opensg Starter Guide 1.2.0
    OpenSG Starter Guide 1.2.0 Generated by Doxygen 1.3-rc2 Wed Mar 19 06:23:28 2003 Contents 1 Introduction 3 1.1 What is OpenSG? . 3 1.2 What is OpenSG not? . 4 1.3 Compilation . 4 1.4 System Structure . 5 1.5 Installation . 6 1.6 Making and executing the test programs . 6 1.7 Making and executing the tutorials . 6 1.8 Extending OpenSG . 6 1.9 Where to get it . 7 1.10 Scene Graphs . 7 2 Base 9 2.1 Base Types . 9 2.2 Log . 9 2.3 Time & Date . 10 2.4 Math . 10 2.5 System . 11 2.6 Fields . 11 2.7 Creating New Field Types . 12 2.8 Base Functors . 13 2.9 Socket . 13 2.10 StringConversion . 16 3 Fields & Field Containers 17 3.1 Creating a FieldContainer instance . 17 3.2 Reference counting . 17 3.3 Manipulation . 18 3.4 FieldContainer attachments . 18 ii CONTENTS 3.5 Data separation & Thread safety . 18 4 Image 19 5 Nodes & NodeCores 21 6 Groups 23 6.1 Group . 23 6.2 Switch . 23 6.3 Transform . 23 6.4 ComponentTransform . 23 6.5 DistanceLOD . 24 6.6 Lights . 24 7 Drawables 27 7.1 Base Drawables . 27 7.2 Geometry . 27 7.3 Slices . 35 7.4 Particles . 35 8 State Handling 37 8.1 BlendChunk . 39 8.2 ClipPlaneChunk . 39 8.3 CubeTextureChunk . 39 8.4 LightChunk . 39 8.5 LineChunk . 39 8.6 PointChunk . 39 8.7 MaterialChunk . 40 8.8 PolygonChunk . 40 8.9 RegisterCombinersChunk .
    [Show full text]
  • A Survey of Technologies for Building Collaborative Virtual Environments
    The International Journal of Virtual Reality, 2009, 8(1):53-66 53 A Survey of Technologies for Building Collaborative Virtual Environments Timothy E. Wright and Greg Madey Department of Computer Science & Engineering, University of Notre Dame, United States Whereas desktop virtual reality (desktop-VR) typically uses Abstract—What viable technologies exist to enable the nothing more than a keyboard, mouse, and monitor, a Cave development of so-called desktop virtual reality (desktop-VR) Automated Virtual Environment (CAVE) might include several applications? Specifically, which of these are active and capable display walls, video projectors, a haptic input device (e.g., a of helping us to engineer a collaborative, virtual environment “wand” to provide touch capabilities), and multidimensional (CVE)? A review of the literature and numerous project websites indicates an array of both overlapping and disparate approaches sound. The computing platforms to drive these systems also to this problem. In this paper, we review and perform a risk differ: desktop-VR requires a workstation-class computer, assessment of 16 prominent desktop-VR technologies (some mainstream OS, and VR libraries, while a CAVE often runs on building-blocks, some entire platforms) in an effort to determine a multi-node cluster of servers with specialized VR libraries the most efficacious tool or tools for constructing a CVE. and drivers. At first, this may seem reasonable: different levels of immersion require different hardware and software. Index Terms—Collaborative Virtual Environment, Desktop However, the same problems are being solved by both the Virtual Reality, VRML, X3D. desktop-VR and CAVE systems, with specific issues including the management and display of a three dimensional I.
    [Show full text]
  • First 6-Monthly EPOCH Pipeline Description
    IST-2002- 507382 EPOCH Excellence in Processing Open Cultural Heritage Network of Excellence Information Society Technologies D3.3.2: 2nd 6-monthly EPOCH Pipeline Description Due date of deliverable: 29 April 2005 Actual submission date: 27 April 2005 Start date of project: 15 March 2004 Duration: 4 Years University of Leuven Project co-funded by the European Commission within the Sixth Framework Programme (2002-2006) Dissemination Level PU Public X PP Restricted to other programme participants (including the Commission Services) RE Restricted to a group specified by the consortium (including the Commission Services) CO Confidential, only for members of the consortium (including the Commission Services) D3.3.2: SECOND 6-MONTHLY EPOCH PIPELINE DESCRIPTION Table of Contents Work package 3, activity 3.3: Common infrastructure .......................................... 3 Objectives.............................................................................................................................. 3 Description of work............................................................................................................... 3 Input from other EPOCH teams............................................................................... 5 Input from the Stakeholder Needs team ................................................................................ 5 Input from the Standards team .............................................................................................. 5 From the Pipeline to a Common Infrastructure ....................................................
    [Show full text]
  • A Scalable Spatial Sound Library for Opensg
    OpenSG Symposium (2003) D. Reiners (Editor) TRIPS – A Scalable Spatial Sound Library for OpenSG Tomas Neumann, Christoph Fünfzig and Dieter Fellner Institute of Computer Graphics, Technical University of Braunschweig, Germany Abstract We present a Sound Library that is scalable on several computers and that brings the idea of a self-organized scenegraph to the Sound Library’s interface. It supports the implementation of audio features right from the start at a high productivity level for rapid prototypes as well as for professional applications in the immersive domain. The system is built on top of OpenSG 12 which offers a high level of functionality for visual applications and research, but does not come with audio support. We show and compare the effort to implement audio in an OpenSG application with and without TRIPS. Today’s audio systems only accept raw 3D-coordinates and are limited to run on the same computer and the same operating system than the application runs on. Breaking these constraints could give developers more freedom and ease to add high-quality spatial sound to their software. Therefore, users benefit from the promising potential OpenSG offers. Categories and Subject Descriptors (according to ACM CCS): H.5.1 [Multimedia Information Systems]: Audio input/output H.5.5 [Sound and Music Computing]: Systems I.3.7 [Computer Graphics]: Virtual Reality Keywords: High quality spatial sound, 3D-audio, cluster system, sound API, FieldContainer, rapid prototyping, game engine, immersive system 1. Introduction tions and computer hardware rapidly developed in the last 10 years, so did the sound and multimedia field. But in a Unix- Current developments on VR applications in general and of based environment nearly all sound engines lack the ability 12 OpenSG applications in particular demand a solution for to benefit from today’s consumer soundboards, which have 3D audio support, as positional sound is more important heavily progressed over the last years as well.
    [Show full text]
  • Scene Graph Rendering
    Scene Graph Rendering Dirk Reiners OpenSG Forum [email protected] March 5, 2002 This part of the course will introduce you to scenegraph systems for rendering. Scene graphs can help simplifying application development and making optimal use of the available graphics hardware. It is assumed that you have some basic knowledge about 3D computer graphics. If the words polygon, directional light source and texture mean nothing to you, you might have problems following the course. The ideal entry level would be having written some programs using OpenGL, or having read the “OpenGL Programming Guide”[1] or “Computer Graphics”[2]. The first section describes the basic structure of a scenegraph and the difference to a standard OpenGL program. As there are a large number of scene graphs around, Open Source and commercial, section 2 gives a short overview of the most commonly used ones. The remainder of this chapter will describe general concepts applying to most scene graphs and use OpenSG as a specific example. The next two sections describe the general node structure and how the graph is traversed. The most important leaf node, the Geometry, is described in section 5. The other of specifying the displayed geometry, the state of transformation and materials etc. is covered in section 6. Some more scene graph specific functionality is hidden inside other node types, as described in sec. 7. To minimize the memory footprint of a scene graph, data can be shared in different ways, sec. 8 gives details. Sec. 9 touches on the importance of multi- threading and what is done in scene graphs to support it.
    [Show full text]
  • The Vjavatar Library: a Toolkit for Development of Avatars in Virtual Reality Applications
    Iowa State University Capstones, Theses and Retrospective Theses and Dissertations Dissertations 1-1-2004 The vjAvatar library: a toolkit for development of avatars in virtual reality applications Justin Allen Hare Iowa State University Follow this and additional works at: https://lib.dr.iastate.edu/rtd Recommended Citation Hare, Justin Allen, "The vjAvatar library: a toolkit for development of avatars in virtual reality applications" (2004). Retrospective Theses and Dissertations. 20575. https://lib.dr.iastate.edu/rtd/20575 This Thesis is brought to you for free and open access by the Iowa State University Capstones, Theses and Dissertations at Iowa State University Digital Repository. It has been accepted for inclusion in Retrospective Theses and Dissertations by an authorized administrator of Iowa State University Digital Repository. For more information, please contact [email protected]. The vjAvatar Library: A toolkit for development of avatars in virtual reality applications by Justin Allen Hare A thesis submitted to the graduate faculty in partial fulfillment of the requirements for the degree of MASTER OF SCIENCE Major: Computer Science Program of Study Committee: Carolina Cruz-Neira, Major Professor Yan-bin Jia Adrian Sannier James Oliver Iowa State University Ames, Iowa 2004 Copyright ©Justin Allen Hare, 2004. All rights reserved 11 Graduate College Iowa State University This is to certify that the Master's thesis of Justin Allen Hare has. met the thesis requirements of Iowa State University Signatures have been redacted for
    [Show full text]
  • The Opensg Forum White Paper
    opensg.org The OpenSG Forum White Paper The Challenge Especially with industrially used VR systems, there is a great demand for a rendering kernel system of the following architecture: Large Models and CAD-Data Integration Scenegraph OpenGL IRIX, Windows, LINUX, etc. On the lowest level, several operating systems (above all SGI IRIX, Windows systems, and Linux) are supported. Above, there is a scenegraph layer guaranteeing the performance and quality that are necessary for VR applications (including the support of parallel processes and of several graphics sub-systems). The top level, finally, consists of all routines for the handling of complex models and for the integration of CAD data. The Starting Point Today, there exist some rendering systems, but they meet the above demands only partially. Moreover, in the past, several systems and extensions have been announced and not realized, or even delivered in prototype state and not advanced any further. Great hopes were placed in the rendering API "Fahrenheit" that was announced by SGI in cooperation with Microsoft. In August 1999, even those contracts were canceled. The current situation is equally depressing and hopeless for VR developers, VR users, and the VR research: For several years now, most VR developers have been prepared to realize a re- design on a new rendering platform, in order to be able to support all the above requirements in their products. The lacking availability was extremely obstructive to the development of VR in general (e.g., missing connections to the industry’s CAx environments, etc.). Developments that had already been started proved as bad investments, and a decision for the integration of a new rendering system cannot be made at the moment.
    [Show full text]
  • DELIVERABLE REPORT Doc
    DELIVERABLE REPORT Doc. Identifier: D. 4.1 Date: 28-11-2011 DELIVERABLE REPORT D 4.1 Document identifier: V-Must.net - D 4.1 Due Date of Delivery to EC End of Month 9 – 30 October 2011 Actual Date of Delivery to EC 28/11/2011 Document date: 22/11/2011 Available platform components and platform Deliverable Title: specification Work package: WP 4 Lead Beneficiary: CNR Other Beneficiaries FHG, CULTNAT, KCL, ULUND, INRIA, SEAV, CINECA, VISDIM, ETF Authors: [Contributors] Massimiliano Corsini, Roberto Scopigno, Luigi Calori, Holger Graf, Sorin Hermon, Daniel Pletinckx Document status: Document link: http://www.v-must.net/media_pubs/documents Grant Agreement 270404 CNR Public 1 /61 DELIVERABLE REPORT Doc. Identifier: D. 4.1 Date: 28-11-2011 Changes Version Date Comment Authors 1.1 02/11/2011 First version 1.2 16/11/2011 Second version 1.3 22/11/2011 Third version Copyright notice: Copyright © V-Must.net. For more information on V-Must.net, its partners and contributors please see http://www.v-must.net/ The information contained in this document reflects only the author's views and the Community is not liable for any use that may be made of the information contained therein. Grant Agreement 270404 CNR Public 2 /61 DELIVERABLE REPORT Doc. Identifier: D. 4.1 Date: 28-11-2011 Table of Contents 1. Executive summary.............................................................................................................5 2. Introduction.........................................................................................................................5
    [Show full text]
  • Chapter 4 Procedural Modeling of Complex Objects Using The
    Chapter 4 Procedural Modeling of Complex Objects using the GPU 1 Chapter 4. Procedural Modeling of Complex Objects using the GPU 4.1 Introduction Procedural modeling involves the use of functional specifications of objects to gen- erate complex geometry from a relatively compact set of descriptors, allowing artists to build detailed models without manually specifying each element, with applications to organic modeling, natural scene generation, city layout, and building architecture. Early systems such as those by Lindenmayer and Stiny used grammatic rules to construct new geometric elements. A key feature of procedural models is that geometry is generated as needed by the grammar or language [Lindenmayer, 1968] [Stiny and Gips, 1971]. This may be contrast with scene graphs, which were historically used in systems such as IRIS Performer to group static geometric primitives for efficient hardware rendering of large scenes [Strauss, 1993] [Rohlf and Helman, 1994]. Scene graphs take advan- tage of persistent GPU buffers, and reordering of graphics state switches, to render these scenes in real time, and are ideally suited to static geometry with few scene changes. A more recent trend is to transmit simplified geometry to graphics hard- ware and allow the GPU to dynamically build detailed models without returning to the CPU. This technique has been successfully applied to mesh refinement, charac- ter skinning, displacement mapping, and terrain rendering [Lorenz and Dollner, 2008] [Rhee et al., 2006] [Szirmay-Kalos and Umenhoffer, 2006]. However, it remains an open question as to how to best take advantage of the GPU for generic procedural modeling. Several early procedural systems developed from the study of nature.
    [Show full text]
  • Opensg: a Scene Graph System for Flexible and Efficient Realtime
    OpenSG: A Scene Graph System for Flexible and Efficient Realtime Rendering for Virtual and Augmented Reality Applications Vom Fachbereich Informatik der Technischen Universität Darmstadt genehmigte DISSERTATION zur Erlangung des akademischen Grades eines Doktor-Ingenieur (Dr.-Ing.) von Dipl.-Inform. Dirk Reiners aus Bergisch Gladbach Referenten der Arbeit: Prof. Dr. José L. Encarnação Prof. Dr. Stefan Müller Tag der Einreichung: 5. Mai 2002 Tag der mündlichen Prüfung: 21. Juni 2002 D17 Darmstädter Dissertationen 2002 Acknowledgements This work wouldn’t have been possible without the continuous support and encour- agement from a number of people. I would like to thank Prof. Dr. José L. Encarnação for creating and maintaining an en- vironment that encourages experimentation and free research without letting people drift too far off track, but keeping them focussed on what needs to be done. My department head Stefan Müller was always there for me and kept believing in me even when I didn’t. Which can be very annoying, but sometimes is just what’s necessary. Many ideas were formed and formalized in intensive discussions with the rest of the OpenSG dream team, Gerrit Voss and Johannes Behr, without whom many mistakes would have been made and many good aproaches ignored. Many ideas were fine-tuned in discussion with other members of the graphics and later OpenSG community. Allen Bierbaum and David Blythe were instrumental in getting everything started. All my colleagues from the A4 department were helpful in making even the hard times fun and enjoyable. Their influence on the overall work environment and general motivation can not be overestimated.
    [Show full text]