Bibliography.Pdf

Total Page:16

File Type:pdf, Size:1020Kb

Bibliography.Pdf Bibliography Thomas Annen, Jan Kautz, Frédo Durand, and Hans-Peter Seidel. Spherical harmonic gradients for mid-range illumination. In Alexander Keller and Henrik Wann Jensen, editors, Eurographics Symposium on Rendering, pages 331–336, Norrköping, Sweden, 2004. Eurographics Association. ISBN 3-905673-12-6. 35,72 Arthur Appel. Some techniques for shading machine renderings of solids. In AFIPS 1968 Spring Joint Computer Conference, volume 32, pages 37–45, 1968. 20 Okan Arikan, David A. Forsyth, and James F. O’Brien. Fast and detailed approximate global illumination by irradiance decomposition. In ACM Transactions on Graphics (Proceedings of SIGGRAPH 2005), pages 1108–1114. ACM Press, July 2005. doi: 10.1145/1073204.1073319. URL http://doi.acm.org/10.1145/1073204.1073319. 47, 48, 52 James Arvo. The irradiance jacobian for partially occluded polyhedral sources. In Computer Graphics (Proceedings of SIGGRAPH 94), pages 343–350. ACM Press, 1994. ISBN 0-89791-667-0. doi: 10.1145/192161.192250. URL http://dx.doi.org/10.1145/192161.192250. 72, 102 Neeta Bhate. Application of rapid hierarchical radiosity to participating media. In Proceedings of ATARV-93: Advanced Techniques in Animation, Rendering, and Visualization, pages 43–53, Ankara, Turkey, July 1993. Bilkent University. 69 Neeta Bhate and A. Tokuta. Photorealistic volume rendering of media with directional scattering. In Alan Chalmers, Derek Paddon, and François Sillion, editors, Rendering Techniques ’92, Eurographics, pages 227–246. Consolidation Express Bristol, 1992. 69 Miguel A. Blanco, M. Florez, and M. Bermejo. Evaluation of the rotation matrices in the basis of real spherical harmonics. Journal Molecular Structure (Theochem), 419:19–27, December 1997. doi: 10.1016/S0166-1280(97)00185-1. URL http://dx.doi.org/10.1016/S0166-1280(97) 00185-1. 34, 176 Philippe Blasi, Bertrand Le Saec,˜ and Christophe Schlick. A rendering algorithm for discrete volume density objects. Computer Graphics Forum (Eurographics ’93), 12(3):201–210, 1993. 64 James F.Blinn. Light reflection functions for simulation of clouds and dusty surfaces. Computer Graphics (Proceedings of SIGGRAPH 82), 16(3):21–29, 1982. ISSN 0097-8930. doi: 10.1145/ 965145.801255. URL http://doi.acm.org/10.1145/965145.801255.3,68 188 189 Pierre Bouguer. Essai d’optique sur la gradation de la lumiere. Jombert, Paris, reprinted in: Les maıtres de la pensee scientifique, Paris, 1729. 59 Leo Breiman, William Meisel, and Edward Purcell. Variable kernel estimates of multivariate densities. Technometrics, 19(2):135–144, May 1977. 139, 186 Brian Cabral, Nelson Max, and Rebecca Springmeyer. Bidirectional reflection functions from surface bump maps. In Computer Graphics (Proceedings of SIGGRAPH 87), pages 273–281. ACM Press, July 1987. ISBN 0-89791-227-6. doi: 10.1145/37401.37434. URL http://doi.acm.org/ 10.1145/37401.37434. 167 Eva Cerezo, Frederic Pérez, Xavier Pueyo, Francisco J. Seron, and François X. Sillion. A survey on participating media rendering techniques. The Visual Computer, 21(5):303–328, June 2005. doi: 10.1007/s00371-005-0287-1. URL http://dx.doi.org/10.1007/s00371-005-0287-1. 68 Subrahmanyan Chandrasekhar. Radiative Transfer. Dover Publications, New York, 1960. 66,68, 69 Shenchang Eric Chen, Holly E. Rushmeier, Gavin Miller, and Douglass Turner. A progressive multi-pass method for global illumination. In Thomas W. Sederberg, editor, Computer Graphics (Proceedings of SIGGRAPH 91), volume 25, pages 165–174. ACM Press, July 1991. ISBN 0-89791- 436-8. doi: 10.1145/122718.122737. URL http://doi.acm.org/10.1145/122718.122737. 21 Kenneth Chiu, Peter Shirley, and Changyaw Wang. Multi-jittered Sampling, pages 370–374. Graphics Gems Series. Academic Press Professional, Inc., San Diego, CA, USA, 1994. ISBN 0-12-336155-9. 162 Cheol Ho Choi, Joseph Ivanic, Mark S. Gordon, and Klaus Ruedenberg. Rapid and stable de- termination of rotation matrices between spherical harmonics by direct recursion. The Journal of Chemical Physics, 111(19):8825–8831, 1999. doi: 10.1063/1.480229. URL http: //dx.doi.org/10.1063/1.480229. 34, 176 Per H. Christensen. Faster photon map global illumination. Journal of Graphics Tools, 4(3):1–10, 1999. 97 Per H. Christensen. Adjoints and importance in rendering: an overview. IEEE Transactions on Visualization and Computer Graphics, 9(3):329–340, July 2003. ISSN 1077-2626. doi: 10.1109/ TVCG.2003.1207441. URL http://dx.doi.org/10.1109/TVCG.2003.1207441. 122 Per H. Christensen, Eric J. Stollnitz, and David H. Salesin. Global illumination of glossy envi- ronments using wavelets and importance. ACM Transactions on Graphics, 15(1):37–71, 1996. ISSN 0730-0301. doi: 10.1145/226150.226153. URL http://doi.acm.org/10.1145/226150. 226153. 20 Michael F. Cohen and Donald P. Greenberg. The hemi-cube; a radiosity solution for complex environments. Computer Graphics (Proceedings of SIGGRAPH 85), 19(3):31–40, August 1985. ISSN 0097-8930. doi: 10.1145/325165.325171. URL http://doi.acm.org/10.1145/325165. 325171. 20 Edward Uhler Condon and George Shortley. The Theory of Atomic Spectra. Cambridge University Press, 1951. 169 190 Robert L. Cook. Stochastic sampling in computer graphics. ACM Transactions on Graphics, 5(1): 51–72, January 1986. ISSN 0730-0301. doi: 10.1145/7529.8927. URL http://doi.acm.org/10. 1145/7529.8927. 21, 162 Robert L. Cook, Thomas Porter, and Loren Carpenter. Distributed ray tracing. Computer Graphics (Proceedings of SIGGRAPH 84), pages 137–145, July 1984. ISSN 0097-8930. doi: 10.1145/964965. 808590. URL http://doi.acm.org/10.1145/964965.808590. 21 Leonardo Da Vinci. A Treatise on Painting. 1651.1 Mark A. Z. Dippé and Erling Henry Wold. Antialiasing through stochastic sampling. Computer Graphics (Proceedings of SIGGRAPH 85), 19(3):69–78, 1985. ISSN 0097-8930. doi: 10.1145/ 325165.325182. URL http://doi.acm.org/10.1145/325165.325182. 164 Frédo Durand, Nicolas Holzschuch, Cyril Soler, Eric Chan, and François Sillion. A frequency analysis of light transport. ACM Transactions on Graphics (Proceedings of SIGGRAPH 2005), 24 (3):1115–1126, August 2005. ISSN 0730-0301. doi: http://doi.acm.org/10.1145/1073204.1073320. URL 10.1145/1073204.1073320. 72 Philip Dutré, Philippe Bekaert, and Kavita Bala. Advanced Global Illumination. AK Peters, Ltd., second edition, 2006. 149, 178 David S. Ebert, Kenton F.Musgrave, Darwyn Peachey, Ken Perlin, and Steven Worley. Texturing & Modeling: A Procedural Approach, Third Edition (The Morgan Kaufmann Series in Computer Graphics). Morgan Kaufmann, December 2002. ISBN 1558608486. 94 Jeppe Revall Frisvad, Niels Jørgen Christensen, and Henrik Wann Jensen. Computing the scattering properties of participating media using lorenz-mie theory. In ACM Transactions on Graphics (Proceedings of SIGGRAPH 2007), page 60, New York, NY, USA, 2007. ACM. doi: 10.1145/1275808. 1276452. URL http://dx.doi.org/10.1145/1275808.1276452. 66 Pascal Gautron, Jaroslav Kˇrivánek, Sumanta N. Pattanaik, and Kadi Bouatouch. A novel hemi- spherical basis for accurate and efficient rendering. In Alexander Keller and Henrik Wann Jensen, editors, Eurographics Symposium on Rendering, pages 321–330, Norrköping, Sweden, 2004. Eurographics Association. ISBN 3-905673-12-6. 33,34 Andrew S. Glassner. Principles of Digital Image Synthesis. Morgan Kaufmann, 1995. 89 Samuel Glasstone and Alexander Sesonske. Nuclear Reactor Engineering. Van Nostrand Company, 1955.3 Cindy M. Goral, Kenneth E. Torrance, Donald P. Greenberg, and Bennett Battaile. Modeling the interaction of light between diffuse surfaces. In Computer Graphics (Proceedings of ACM SIGGEAPH 84), pages 213–222, New York, NY, USA, 1984. ACM Press. ISBN 0-89791-138-5. doi: 10.1145/800031.808601. URL http://doi.acm.org/10.1145/800031.808601. 20 Steven J. Gortler, Peter Schröder, Michael F. Cohen, and Pat Hanrahan. Wavelet radiosity. In Computer Graphics (Proceedings of SIGGRAPH 93), pages 221–230, New York, NY, USA, 1993. ACM. ISBN 0-89791-601-8. doi: 10.1145/166117.166146. URL http://doi.acm.org/10.1145/ 166117.166146. 20 191 Robin Green. Spherical harmonic lighting: The gritty details. Archives of the Game Developers Conference, March 2003. 167, 176 Pat Hanrahan and Wolfgang Krueger. Reflection from layered surfaces due to subsurface scattering. In Computer Graphics (Proceedings of SIGGRAPH 93), pages 165–174, New York, NY, USA, 1993. ACM. ISBN 0-89791-601-8. doi: 10.1145/166117.166139. URL http://dx.doi.org/10.1145/ 166117.166139. 70 Pat Hanrahan, David Salzman, and Larry Aupperle. A rapid hierarchical radiosity algorithm. Computer Graphics (Proceedings of SIGGRAPH 91), 25(4):197–206, 1991. ISSN 0097-8930. doi: 10.1145/127719.122740. URL http://doi.acm.org/10.1145/127719.122740. 20 Louis George Henyey and Jesse Leonard Greenstein. Diffuse radiation in the galaxy. Astrophysics, 93:70–83, 1941. 63 Nicolas Holzschuch and François Sillion. Accurate computation of the radiosity gradient with con- stant and linear emitters. In Patrick M. Hanrahan and Werner Purgathofer, editors, Rendering Techniques ’95, Eurographics, pages 186–195. Springer-Verlag Wien New York, 1995. 72 Nicolas Holzschuch and François Sillion. An exhaustive error-bounding algorithm for hierarchical radiosity. Computer Graphics Forum, 17(4), 1998. 72 Hoyt C. Hottel and Adel F.Sarofim. Radiative Transfer. McGraw Hill, New York, 1967. 69 Homan Igehy. Tracing ray differentials. In Computer Graphics (Proceedings of SIGGRAPH 99), pages 179–186, New York, NY, USA, 1999. ACM Press/Addison-Wesley Publishing Co. ISBN 0-201- 48560-5. doi: 10.1145/311535.311555. URL http://doi.acm.org/10.1145/311535.311555. 72 David S. Immel, Michael F.Cohen, and Donald P.Greenberg. A radiosity method for non-diffuse environments. Computer Graphics (Proceedings of SIGGRAPH 86), 20(4):133–142, 1986. ISSN 0097-8930. doi: 10.1145/15886.15901. URL http://doi.acm.org/10.1145/15886.15901.20 Joseph Ivanic and Klaus Ruedenberg. Rotation matrices for real spherical harmonics. direct determination by recursion. Journal of Physical Chemistry, 100(15):6342–6347, 1996. ISSN 0022-3654.
Recommended publications
  • Efficient Ray Tracing of Volume Data
    Efficient Ray Tracing of Volume Data TR88-029 June 1988 Marc Levay The University of North Carolina at Chapel Hill Department of Computer Science CB#3175, Sitterson Hall Chapel Hill, NC 27599-3175 UNC is an Equal Opportunity/ Affirmative Action Institution. 1 Efficient Ray Tracing of Volume Data Marc Levoy June, 1988 Computer Science Department University of North Carolina Chapel Hill, NC 27599 Abstract Volume rendering is a technique for visualizing sampled scalar functions of three spatial dimen­ sions without fitting geometric primitives to the data. Images are generated by computing 2-D pro­ jections of a colored semi-transparent volume, where the color and opacity at each point is derived from the data using local operators. Since all voxels participate in the generation of each image, rendering time grows linearly with the size of the dataset. This paper presents a front-to-back image-order volume rendering algorithm and discusses two methods for improving its performance. The first method employs a pyramid of binary volumes to encode coherence present in the data, and the second method uses an opacity threshold to adaptively terminate ray tracing. These methods exhibit a lower complexity than brute-force algorithms, although the actual time saved depends on the data. Examples from two applications are given: medical imaging and molecular graphics. 1. Introduction The increasing availability of powerful graphics workstations in the scientific and computing communities has catalyzed the development of new methods for visualizing discrete multidimen­ sional data. In this paper, we address the problem of visualizing sampled scalar functions of three spatial dimensions, henceforth referred to as volume data.
    [Show full text]
  • Volume Rendering on Scalable Shared-Memory MIMD Architectures
    Volume Rendering on Scalable Shared-Memory MIMD Architectures Jason Nieh and Marc Levoy Computer Systems Laboratory Stanford University Abstract time. Examples of these optimizations are hierarchical opacity enumeration, early ray termination, and spatially adaptive image Volume rendering is a useful visualization technique for under- sampling [13, 14, 16, 181. Optimized ray tracing is the fastest standing the large amounts of data generated in a variety of scien- known sequential algorithm for volume rendering. Nevertheless, tific disciplines. Routine use of this technique is currently limited the computational requirements of the fastest sequential algorithm by its computational expense. We have designed a parallel volume still preclude real-time volume rendering on the fastest computers rendering algorithm for MIMD architectures based on ray tracing today. and a novel task queue image partitioning technique. The combi- Parallel machines offer the computational power to achieve real- nation of ray tracing and MIMD architectures allows us to employ time volume rendering on usefully large datasets. This paper con- algorithmic optimizations such as hierarchical opacity enumera- siders how parallel computing can be brought to bear in reducing tion, early ray termination, and adaptive image sampling. The use the computation time of volume rendering. We have designed of task queue image partitioning makes these optimizations effi- and implemented an optimized volume ray tracing algorithm that cient in aparallel framework. We have implemented our algorithm employs a novel task queue image partitioning technique on the on the Stanford DASH Multiprocessor, a scalable shared-memory Stanford DASH Multiprocessor [lo], a scalable shared-memory MIMD machine. Its single address-space and coherent caches pro- MIMD (Multiple Instruction, Multiple Data) machine consisting vide programming ease and good performance for our algorithm.
    [Show full text]
  • Computational Photography
    Computational Photography George Legrady © 2020 Experimental Visualization Lab Media Arts & Technology University of California, Santa Barbara Definition of Computational Photography • An R & D development in the mid 2000s • Computational photography refers to digital image-capture cameras where computers are integrated into the image-capture process within the camera • Examples of computational photography results include in-camera computation of digital panoramas,[6] high-dynamic-range images, light- field cameras, and other unconventional optics • Correlating one’s image with others through the internet (Photosynth) • Sensing in other electromagnetic spectrum • 3 Leading labs: Marc Levoy, Stanford (LightField, FrankenCamera, etc.) Ramesh Raskar, Camera Culture, MIT Media Lab Shree Nayar, CV Lab, Columbia University https://en.wikipedia.org/wiki/Computational_photography Definition of Computational Photography Sensor System Opticcs Software Processing Image From Shree K. Nayar, Columbia University From Shree K. Nayar, Columbia University Shree K. Nayar, Director (lab description video) • Computer Vision Laboratory • Lab focuses on vision sensors; physics based models for vision; algorithms for scene interpretation • Digital imaging, machine vision, robotics, human- machine interfaces, computer graphics and displays [URL] From Shree K. Nayar, Columbia University • Refraction (lens / dioptrics) and reflection (mirror / catoptrics) are combined in an optical system, (Used in search lights, headlamps, optical telescopes, microscopes, and telephoto lenses.) • Catadioptric Camera for 360 degree Imaging • Reflector | Recorded Image | Unwrapped [dance video] • 360 degree cameras [Various projects] Marc Levoy, Computer Science Lab, Stanford George Legrady Director, Experimental Visualization Lab Media Arts & Technology University of California, Santa Barbara http://vislab.ucsb.edu http://georgelegrady.com Marc Levoy, Computer Science Lab, Stanford • Light fields were introduced into computer graphics in 1996 by Marc Levoy and Pat Hanrahan.
    [Show full text]
  • Computational Photography Research at Stanford
    Computational photography research at Stanford January 2012 Marc Levoy Computer Science Department Stanford University Outline ! Light field photography ! FCam and the Stanford Frankencamera ! User interfaces for computational photography ! Ongoing projects • FrankenSLR • burst-mode photography • languages and architectures for programmable cameras • computational videography and cinematography • multi-bucket pixels 2 ! Marc Levoy Light field photography using a handheld plenoptic camera Ren Ng, Marc Levoy, Mathieu Brédif, Gene Duval, Mark Horowitz and Pat Hanrahan (Proc. SIGGRAPH 2005 and TR 2005-02) Light field photography [Ng SIGGRAPH 2005] ! Marc Levoy Prototype camera Contax medium format camera Kodak 16-megapixel sensor Adaptive Optics microlens array 125µ square-sided microlenses 4000 ! 4000 pixels ÷ 292 ! 292 lenses = 14 ! 14 pixels per lens Digital refocusing ! ! ! 2010 Marc Levoy Example of digital refocusing ! 2010 Marc Levoy Refocusing portraits ! 2010 Marc Levoy Light field photography (Flash demo) ! 2010 Marc Levoy Commercialization • trades off (excess) spatial resolution for ability to refocus and adjust the perspective • sensor pixels should be made even smaller, subject to the diffraction limit, for example: 36mm ! 24mm ÷ 2.5µ pixels = 266 Mpix 20K ! 13K pixels 4000 ! 2666 pixels ! 20 ! 20 rays per pixel or 2000 ! 1500 pixels ! 3 ! 3 rays per pixel = 27 Mpix ! Marc Levoy Camera 2.0 Marc Levoy Computer Science Department Stanford University The Stanford Frankencameras Frankencamera F2 Nokia N900 “F” ! facilitate research in experimental computational photography ! for students in computational photography courses worldwide ! proving ground for plugins and apps for future cameras 14 ! 2010 Marc Levoy CS 448A - Computational Photography 15 ! 2010 Marc Levoy CS 448 projects M. Culbreth, A. Davis!! ! ! A phone-to-phone 4D light field fax machine J.
    [Show full text]
  • Computational Photography & Cinematography
    Computational photography & cinematography (for a lecture specifically devoted to the Stanford Frankencamera, see http://graphics.stanford.edu/talks/camera20-public-may10-150dpi.pdf) Marc Levoy Computer Science Department Stanford University The future of digital photography ! the megapixel wars are over (and it’s about time) ! computational photography is the next battleground in the camera industry (it’s already starting) 2 ! Marc Levoy Premise: available computing power in cameras is rising faster than megapixels 12 70 9 52.5 6 35 megapixels 3 17.5 billions of pixels / second of pixels billions 0 0 2005 2006 2007 2008 2009 2010 Avg megapixels in new cameras, CAGR = 1.2 NVIDIA GTX texture fill rate, CAGR = 1.8 (CAGR for Moore’s law = 1.5) ! this “headroom” permits more computation per pixel, 3 or more frames per second, or less custom hardware ! Marc Levoy The future of digital photography ! the megapixel wars are over (long overdue) ! computational photography is the next battleground in the camera industry (it’s already starting) ! how will these features appear to consumers? • standard and invisible • standard and visible (and disable-able) • aftermarket plugins and apps for your camera 4 ! Marc Levoy The future of digital photography ! the megapixel wars are over (long overdue) ! computational photography is the next battleground in the camera industry (it’s already starting) ! how will these features appear to consumers? • standard and invisible • standard and visible (and disable-able) • aftermarket plugins and apps for your camera
    [Show full text]
  • Uva-DARE (Digital Academic Repository)
    UvA-DARE (Digital Academic Repository) Interactive Exploration in Virtual Environments Belleman, R.G. Publication date 2003 Link to publication Citation for published version (APA): Belleman, R. G. (2003). Interactive Exploration in Virtual Environments. General rights It is not permitted to download or to forward/distribute the text or part of it without the consent of the author(s) and/or copyright holder(s), other than for strictly personal, individual use, unless the work is under an open content license (like Creative Commons). Disclaimer/Complaints regulations If you believe that digital publication of certain material infringes any of your rights or (privacy) interests, please let the Library know, stating your reasons. In case of a legitimate complaint, the Library will make the material inaccessible and/or remove it from the website. Please Ask the Library: https://uba.uva.nl/en/contact, or a letter to: Library of the University of Amsterdam, Secretariat, Singel 425, 1012 WP Amsterdam, The Netherlands. You will be contacted as soon as possible. UvA-DARE is a service provided by the library of the University of Amsterdam (https://dare.uva.nl) Download date:30 Sep 2021 References s [1]] 3rdTech webpage, http://www.3rdtech.com/. [2]] M.J. Ackerman. Accessing the visible human project. D-Lib Magazine: The MagazineMagazine of the Digital Library Forum, October 1995. [3]] H. Afsarmanesh, R.G. Belleman, A.S.Z. Belloum, A. Benabdelkader, J.F.J, vann den Brand, G.B. Eijkel, A. Frenkel, C. Garita, D.L. Groep, R.M.A. Heeren, Z.W.. Hendrikse, L.0. Hertzberger, J.A. Kaandorp, E.C.
    [Show full text]
  • Multi-View Image Fusion
    Multi-view Image Fusion Marc Comino Trinidad1 Ricardo Martin Brualla2 Florian Kainz2 Janne Kontkanen2 1Polytechnic University of Catalonia, 2Google Inputs Outputs Color Transfer Mono Color High-def mono Color High-def color HDR Fusion Color EV+2 Color EV-1 Color EV+2 Color EV-1 Denoised Detail Transfer DSLR Stereo DSLR Low-def stereo High-def stereo Figure 1: We present a method for multi-view image fusion that is a applicable to a variety of scenarios: a higher resolution monochrome image is colorized with a second color image (top row), two color images with different exposures are fused into an HDR lower-noise image (middle row), and a high quality DSLR image is warped to the lower quality stereo views captured by a VR-camera (bottom row). Abstract 1. Introduction In this paper we focus on the problem of fusing multiple We present an end-to-end learned system for fusing mul- misaligned photographs into a chosen target view. Multi- tiple misaligned photographs of the same scene into a cho- view image fusion has become increasingly relevant with sen target view. We demonstrate three use cases: 1) color the recent influx of multi-camera mobile devices. transfer for inferring color for a monochrome view, 2) HDR The form factor of these devices constrains the size of fusion for merging misaligned bracketed exposures, and 3) lenses and sensors, and this limits their light capturing abil- detail transfer for reprojecting a high definition image to ity. Cameras with larger lens apertures and larger pixels the point of view of an affordable VR180-camera.
    [Show full text]
  • Proposal for a SIGGRAPH 2007 Course
    Proposal for a SIGGRAPH 2008 Course: Computational Photography: Advanced Topics Online ID: courses_0040 1 Course Title Computational Photography 2 Category Image-based and 3D Photography techniques - other 3 Course Organizer Ramesh Raskar MIT-Media Lab (preferred) Mitsubishi Electric Research Labs (MERL) 20 Ames St., 201 Broadway Cambridge, MA 02139, USA Cambridge, MA 02139, USA Jack Tumblin EECS Dept, 3-320 Ford Design Center Northwestern University 2133 Sheridan Road Evanston, IL 60208, USA 4 Proposed Length 3.75 Hour 5 Proposed Presentation Venue Regular session room 6 Summary Statement Participants learn about the latest computational methods in digital imaging that overcome the traditional limitations of a camera and enable novel imaging applications. The course provides a practical guide to topics in image capture and manipulation methods for generating compelling pictures for computer graphics and for extracting scene properties for computer vision, with several examples. 7 Names of Lecturers Paul DEBEVEC(University of Southern California, Institute for Creative Technologies (USC-ICT), USA)([email protected]) Ramesh RASKAR (Mitsubishi Electric Research Labs (MERL), USA) ([email protected]) Jack TUMBLIN (Northwestern University, USA) ([email protected]) 8 Course Abstract Abstract Computational photography combines plentiful computing, digital sensors, modern optics, many varieties of actuators, probes and smart lights to escape the limitations of traditional film cameras and enables novel imaging applications. Unbounded dynamic range, variable focus, resolution, and depth of field, hints about shape, reflectance, and lighting, and new interactive forms of photos that are partly snapshots and partly videos, performance capture and interchangeably relighting real and virtual characters are just some of the new applications emerging in Computational Photography.
    [Show full text]
  • Digital Paint Systems: an Anecdotal and Historical Overview
    Digital Paint Systems: An Anecdotal and Historical Overview Alvy Ray Smith The history of digital paint systems derives from many things— chance meetings, coincidences and boredom, artistic license, brilliant researchers, a wealthy benefactor, and, of course, lawsuits. Alvy Ray Smith tells the fascinating story—facts first, then anecdotes—in his own words. This article is based on a talk presented in tal 2D (two-dimensional) and 3D modeling and January 2000 at an evening hosted by the animation, film recording, video editing, and Computer History Museum on Moffatt Field, audio synthesis. An excellent rendering of my near Palo Alto, California. I shared the floor time with Dick Shoup (sounds like “shout,” not with longtime colleague Richard G. “Dick” like “hoop”) in the early days at Xerox PARC Shoup who figures highly in what follows. It is (Palo Alto Research Center) can be found in the also based on a document I submitted to the recent book Dealers of Lightning: Xerox PARC Academy of Motion Picture Arts and Sciences and the Dawn of the Computer Age.4 For other (AMPAS) in 1997 in answer to a call from them PARC references, see Lavendel,5 Pake,6 Perry,7 for information about early paint programs and and Smith.8 their contribution to the film business.1 The time frame dates from the late 1960s to Definitions the early 1980s, from the beginnings of the A digital paint program and a digital paint technology of digital painting up to the first system are distinguished by their functions. A consumer products that implemented it.
    [Show full text]
  • Acquisition and Representation of Material Appearance for Editing
    Acquisition and Representation of Material Appearance for Editing and Rendering Jason Davis Lawrence A Dissertation Presented to the Faculty of Princeton University in Candidacy for the Degree of Doctor of Philosophy Recommended for Acceptance By the Department of Computer Science September 2006 c Copyright by Jason Davis Lawrence, 2006. All rights reserved. Abstract Providing computer models that accurately characterize the appearance of a wide class of materials is of great interest to both the computer graphics and computer vision communities. The last ten years has witnessed a surge in techniques for measuring the optical properties of physical materials. As compared to conventional techniques that rely on hand-tuning parametric light reflectance functions, a data-driven approach is better suited for representing complex real-world appearance. However, incorporating these representations into existing rendering algorithms and a practical production pipeline has remained an open research problem. One common approach has been to fit the parameters of an analytic reflectance function to measured appearance data. This has the benefit of providing significant compression ratios and these analytic models are already fully integrated into modern rendering algorithms. However, this approach can lead to significant approximation errors for many materials and it requires computationally expensive and numerically unstable non-linear optimization. An alternative approach is to compress these datasets, using algorithms such as Principal Component Analysis, wavelet compression or matrix factorization. Although these techniques provide an accurate and compact representation, they do have several drawbacks. In particular, existing methods do not enable efficient importance sampling for measured materials (and even some complex analytic models) in the context of physically-based rendering systems.
    [Show full text]
  • Handheld Multi-Frame Super-Resolution
    Handheld Multi-Frame Super-Resolution BARTLOMIEJ WRONSKI, IGNACIO GARCIA-DORADO, MANFRED ERNST, DAMIEN KELLY, MICHAEL KRAININ, CHIA-KAI LIANG, MARC LEVOY, and PEYMAN MILANFAR, Google Research Fig. 1. We present a multi-frame super-resolution algorithm that supplants the need for demosaicing in a camera pipeline by merging a burst of raw images. We show a comparison to a method that merges frames containing the same-color channels together first, and is then followed by demosaicing (top). By contrast, our method (bottom) creates the full RGB directly from a burst of raw images. This burst was captured with a hand-held mobile phone and processed on device. Note in the third (red) inset that the demosaiced result exhibits aliasing (Moiré), while our result takes advantage of this aliasing, which changes on every frame in the burst, to produce a merged result in which the aliasing is gone but the cloth texture becomes visible. Compared to DSLR cameras, smartphone cameras have smaller sensors, CCS Concepts: • Computing methodologies → Computational pho- which limits their spatial resolution; smaller apertures, which limits their tography; Image processing. light gathering ability; and smaller pixels, which reduces their signal-to- Additional Key Words and Phrases: computational photography, super- noise ratio. The use of color filter arrays (CFAs) requires demosaicing, which further degrades resolution. In this paper, we supplant the use of traditional resolution, image processing, photography demosaicing in single-frame and burst photography pipelines with a multi- ACM Reference Format: frame super-resolution algorithm that creates a complete RGB image directly Bartlomiej Wronski, Ignacio Garcia-Dorado, Manfred Ernst, Damien Kelly, from a burst of CFA raw images.
    [Show full text]
  • Proceedings August 3-8, 1997 Papers Chair: Turner Whitted Panels Chair: Barbara Mones-Hattal
    Annual Conference Series 1997 SIGGRAPH 97 Conference Proceedings August 3-8, 1997 Papers Chair: Turner Whitted Panels Chair: Barbara Mones-Hattal A Publication of ACM SIGGRAPH Sponsored by the ACM's Special Interest Group on Computer Graphics PROCEEDINGS COMPUTER GRAPHICS Proceedings, Annual Conference Series, 1997 Contents Papers Preface 8 Panels Preface 9 Papers Sessions, Wednesday, 6 August 8:15 - 9:45 SIGGRAPH 97 Keynote Address Steven A. Coons Award for Outstanding Creative Contributions to Computer Graphics 10 1997 ACM SIGGRAPH Computer Graphics Achievement Award 11 10:15 - 12:00 Virtual Reality and Applications Chair: Frederick P. Brooks, Jr. Quantifying Immersion in Virtual Reality 13 Randy Pausen, Dennis Proffitt, George Williams Moving Objects in Space: Exploiting Proprioception In Virtual-Environment Interaction 19 Mark R. Mine, Frederick P. Brooks, Jr, Carlo H. Se'quin Virtual Voyage: Interactive Navigation in the Human Colon 27 Lichan Hong, Shigeru Muraki, Arie Kaufman, Dirk Bartz, Taosong He Interactive Simulation of Fire in Virtual Building Environments 35 Richard Bukowski, Carlo H. Sequin 2:00 - 3:45 Illumination Chair: Eugene Fiume Fitting Virtual Lights For Non-Diffuse Walkthroughs 45 Bruce Walter, Gün Alppay, Eric P. F. Lafortune, Sebastian Femandez, Donald P. Greenberg Instant Radiosity 49 Alexander Keller Interactive Update of Global Illumination Using A Line-Space Hierarchy 57 George Drettakis, Francois Sillion Metropolis Light Transport 65 Eric Veach, Leonidas J. Guibas 4:00 - 5:45 Visibility Chair: Thomas Funkhouser Visibility Culling Using Hierarchical Occlusion Maps 77 Hansong Zhang, Dinesh Manocha, Thomas Hudson, Kenneth E. Hoff III The Visibility Skeleton: A Powerful and Efficient Multi-Purpose Global Visibility Tool 89 Fredo Durand, George Drettakis, Claude Puech Rendering Complex Scenes with Memory-Coherent Ray Tracing 101 Matt Pharr, Craig Kolb, Reid Gershbein, Pat Hanrahan Illustrating Surface Shape in Volume Data via Principal Direction-Driven 3D Line Integral Convolution 109 Victoria L.
    [Show full text]