Jitter-Camera: High Resolution Video from a Low Resolution Detector

Total Page:16

File Type:pdf, Size:1020Kb

Jitter-Camera: High Resolution Video from a Low Resolution Detector Images in this paper are best viewed magnified or printed on a high resolution color printer. Jitter-Camera: High Resolution Video from a Low Resolution Detector Moshe Ben-Ezra, Assaf Zomet, and Shree K. Nayar Computer Science Department Columbia University New York, NY, 10027 E-mail: {moshe, zomet, nayar}@cs.columbia.edu Abstract Video cameras must produce images at a reasonable frame-rate and with a reasonable depth of field. These requirements impose fundamental physical limits on the spatial resolution of the image detector. As a result, current cameras produce videos with a very low resolution. The resolution of videos can be computationally enhanced by moving the camera and applying super-resolution reconstruction algorithms. However, a moving camera introduces motion blur, which limits super-resolution quality. We analyze this effect and derive a theoretical result showing that motion blur has a substantial degrading effect on the performance of super resolution. The conclusion is, that in order to achieve the highest resolution, motion blur should be avoided. Motion blur can be minimized by sampling the space-time volume of the video in a specific manner. We have developed a novel camera, called the ”jitter camera,” that achieves this sampling. By applying an adaptive super-resolution algorithm to the video produced by the jitter camera, we show that resolution can be notably enhanced for stationary or slowly moving objects, while it is improved slightly or left unchanged for objects with fast and complex motions. The end result is a video that has a significantly higher resolution than the captured one. Keywords: Sensors; Jitter Camera; Jitter Video; Super Resolution; Motion Blur; 2 Time Tim e Space Space (a) (b) Figure 1: Conventional video cameras sample the continuous space-time volume at regular time inter- vals and fixed spatial grid locations as shown in (a). The space-time volume can be sampled differently, for example, by varying the location of the sampling grid as shown in (b) to increase the resolution of the video. A moving video only approximates (b) due to motion blur. 1. Why is High-Resolution Video Hard? Improving the spatial resolution of a video camera is different from doing so with a still camera. Merely increasing the number of pixels of the detector reduces the amount of light received by each pixel, and hence increases the noise. With still images, this can be overcome by prolonging the exposure time. In the case of video, however, the exposure time is limited by the desired frame-rate. The amount of light incident on the detector can also be increased by widening the aperture, but with a significant reduction of the depth of field. The spatial resolution of a video detector is therefore limited by the noise level of the detector, the frame-rate (temporal resolution) and the required depth of field1. Our purpose is to make a judicious use of a given detector, that will allow a substantial increase of the video resolution by a resolution-enhancement algorithm. Figure 1 shows a continuous space-time video volume. A slice of this volume at a given time instance corresponds to the image appearing on the image plane of the camera at this time. This volume is sampled both spatially and temporally, where each pixel integrates light over time and space. Conventional video cameras sample the volume in a simple way, as shown in Figure 1(a), with a regular 2D grid of pixels integrating over regular temporal intervals and at fixed spatial locations. An alternative sampling of the space- time volume is shown in Figure 1(b). The 2D grid of pixels integrates over the same temporal intervals, but at different spatial locations. Given a 2D image detector, how should we sample the space-time volume to 1The optical transfer function of the lens also imposes a limit on resolution. In this paper we ignore this limit as it is several orders of magnitudes above the current resolution of video. 3 obtain the highest spatial resolution2? There is a large body of work on resolution enhancement by varying spatial sampling, commonly known as super-resolution reconstruction [4, 5, 7, 9, 13, 17]. Super-resolution algorithms typically assume that a set of displaced images are given as input. With a video camera, this can be achieved by moving the camera while capturing the video. However, the camera’s motion introduces motion blur. This is a key point in this paper: in order to use super-resolution with a conventional video camera, the camera must move, but when the camera moves, it introduces motion blur which reduces resolution. It is well known that an accurate estimation of the motion blur parameters is non-trivial, and requires strong assumptions about the camera motion during integration [2,13,15,19]. In this paper, we show that even when an accurate estimate of the motion blur parameters is available, motion blur has a significant influence on the super-resolution result. We derive a theoretical lower bound, indicating that the expected performance of any super-resolution reconstruction algorithm deteriorates as a function of the motion blur magnitude. The conclusion is that, in order to achieve the highest resolution, motion blur should be avoided. To achieve this, we propose the “jitter camera,” a novel video camera that samples the space-time volume at different locations without introducing motion blur. This is done by instantaneously shifting the detector (e.g. CCD) between temporal integration periods, rather than continuously moving the entire video camera during the integration periods. We have built a jitter camera, and developed an adaptive super-resolution algorithm to handle complex scenes containing multiple moving objects. By applying the algorithm to the video produced by the jitter camera, we show that resolution can be enhanced significantly for stationary or slowly moving objects, while it is improved slightly or left unchanged for objects with fast and complex motions. The end result is a video that has higher resolution than the captured one. 2. How Bad is Motion Blur for Super-resolution? The influence of motion blur on super resolution is well understood when all input images undergo the same motion blur [1, 10]. It becomes more complex when the input images undergo different motion blurs, and details that appear blurred in one image, appear sharp in another image. We address the influence of motion blur for any combination of blur orientations. 2Increasing the temporal resolution [18] is not addressed in this paper. 4 12 10 8 6 4 Lower Bound 2 s(A), Monotone Function of the Volume Solutions 0 0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5 Length of Motion Blur Trajectory (pixels) Figure 2: We measure the super-resolution “hardness” by the volume of plausible high-resolution so- lutions [1]. The volume of solutions is proportional to s(A)n2 where n2 is the high resolution image size. 3 The graphs show the value of s(A) as a function of the length of the motion-blur trajectories {kljk}j=0. We show a large number of graphs computed for different configurations of blur orientations. The thick graph (blue line) is the lower bound of s(A), for any combination of motion blur orientations. In all shown configurations, the motion blur has a significant influence on s(A) and hence on the volume of solutions. The increase in the volume of solutions can explain the increase in reconstruction error in super-resolution shown in Figure 3. Super-resolution algorithms estimate the high resolution image by modeling and inverting the imaging pro- cess. Analyzing the influence of motion blur requires a definition for super-resolution “hardness” or the “invertibility” of the imaging process. We use a linear model for the imaging process [1, 7, 9, 13], where the intensity of a pixel in the input image is presented as a linear combination of the intensities in the unknown high resolution image: ~y = A~x + ~z, (1) where ~x is a vectorization of the unknown discrete high resolution image, ~y is a vectorization of all the input images, and the imaging matrix A encapsulates the camera displacements, blur and decimation [7]. The random variable ~z represents the uncertainty in the measurements due to noise, quantization error and model inaccuracies. 5 Baker and Kanade [1] addressed the invertibility of the imaging process in a noise-free scenario, where ~z represents the quantization error. In this case, each quantized input pixel defines two inequality constraints on the super-resolution solution. The combination of constraints forms a volume of solutions that satisfy all quantization constraints. Baker and Kanade suggest to use the volume of solutions as a measure of uncertainty in the super resolution solution. Their paper [1] shows the benefits in measuring the volume of solutions over the standard matrix conditioning analysis. We measure the influence of motion blur by the volume of solutions. To keep the analysis simple, the following assumptions are made. First, the motion blur in each input image is induced by a constant velocity motion. Different input images may have different motion blur orientations. Second, the optical blur is shift- invariant. Third, the input images are related geometrically by a 2D translation. Fourth, the number of input pixels equals the number of output pixels. Under the last assumption, the dimensionality n2 of ~x equals the dimensionality of ~y. Since the uncertainty due to quantization is an n2-dimensional unit cube, the volume of solutions for a given imaging matrix A can be computed from the absolute value of its determinant 1 vol(A) = . (2) |A| In Appendix A, we derive a simplified expression for |A| as a function of the imaging parameters.
Recommended publications
  • Understanding Jitter and Wander Measurements and Standards Second Edition Contents
    Understanding Jitter and Wander Measurements and Standards Second Edition Contents Page Preface 1. Introduction to Jitter and Wander 1-1 2. Jitter Standards and Applications 2-1 3. Jitter Testing in the Optical Transport Network 3-1 (OTN) 4. Jitter Tolerance Measurements 4-1 5. Jitter Transfer Measurement 5-1 6. Accurate Measurement of Jitter Generation 6-1 7. How Tester Intrinsics and Transients 7-1 affect your Jitter Measurement 8. What 0.172 Doesn’t Tell You 8-1 9. Measuring 100 mUIp-p Jitter Generation with an 0.172 Tester? 9-1 10. Faster Jitter Testing with Simultaneous Filters 10-1 11. Verifying Jitter Generator Performance 11-1 12. An Overview of Wander Measurements 12-1 Acronyms Welcome to the second edition of Agilent Technologies Understanding Jitter and Wander Measurements and Standards booklet, the distillation of over 20 years’ experience and know-how in the field of telecommunications jitter testing. The Telecommunications Networks Test Division of Agilent Technologies (formerly Hewlett-Packard) in Scotland introduced the first jitter measurement instrument in 1982 for PDH rates up to E3 and DS3, followed by one of the first 140 Mb/s jitter testers in 1984. SONET/SDH jitter test capability followed in the 1990s, and recently Agilent introduced one of the first 10 Gb/s Optical Channel jitter test sets for measurements on the new ITU-T G.709 frame structure. Over the years, Agilent has been a significant participant in the development of jitter industry standards, with many contributions to ITU-T O.172/O.173, the standards for jitter test equipment, and Telcordia GR-253/ ITU-T G.783, the standards for operational SONET/SDH network equipment.
    [Show full text]
  • Motion Denoising with Application to Time-Lapse Photography
    IEEE Computer Vision and Pattern Recognition (CVPR), June 2011 Motion Denoising with Application to Time-lapse Photography Michael Rubinstein1 Ce Liu2 Peter Sand Fredo´ Durand1 William T. Freeman1 1MIT CSAIL 2Microsoft Research New England {mrub,sand,fredo,billf}@mit.edu [email protected] Abstract t Motions can occur over both short and long time scales. We introduce motion denoising, which treats short-term x changes as noise, long-term changes as signal, and re- Input renders a video to reveal the underlying long-term events. y We demonstrate motion denoising for time-lapse videos. One of the characteristics of traditional time-lapse imagery t is stylized jerkiness, where short-term changes in the scene x t (Time) appear as small and annoying jitters in the video, often ob- x Motion-denoised fuscating the underlying temporal events of interest. We ap- ply motion denoising for resynthesizing time-lapse videos showing the long-term evolution of a scene with jerky short- term changes removed. We show that existing filtering ap- proaches are often incapable of achieving this task, and present a novel computational approach to denoise motion without explicit motion analysis. We demonstrate promis- InputDisplacement Result ing experimental results on a set of challenging time-lapse Figure 1. A time-lapse video of plants growing (sprouts). XT sequences. slices of the video volumes are shown for the input sequence and for the result of our motion denoising algorithm (top right). The motion-denoised sequence is generated by spatiotemporal rear- 1. Introduction rangement of the pixels in the input sequence (bottom center; spa- tial and temporal displacement on top and bottom respectively, fol- Short-term, random motions can distract from the lowing the color coding in Figure 5).
    [Show full text]
  • DISTRIBUTED RAY TRACING – Some Implementation Notes Multiple Distributed Sampling Jitter - to Break up Patterns
    DISTRIBUTED RAY TRACING – some implementation notes multiple distributed sampling jitter - to break up patterns DRT - theory v. practice brute force - generate multiple rays at every sampling opportunity alternative: for each subsample, randomize at each opportunity DRT COMPONENTS anti-aliasing and motion blur supersampling - in time and space: anti-aliasing jitter sample in time and space: motion blur depth of field - sample lens: blurs shadows – sample light source: soft shadows reflection – sample reflection direction: rough surface transparency – sample transmission direction: translucent surface REPLACE CAMERA MODEL shift from pinhole camera model to lens camera model picture plane at -w, not +w camera position becomes lens center picture plane is behind 'pinhole' negate u, v, w, trace ray from pixel to camera ANTI-ALIASING: ORGANIZING subpixel samples Options 1. Do each subsample in raster order 2. do each pixel in raster order, do each subsample in raster order 3. do each pixel in raster order, do all subsamples in temporal order 4. keep framebuffer, do all subsamples in temporal order SPATIAL JITTERING for each pixel 200x200, i,j for each subpixel sample 4x4 s,t JITTERED SAMPLE jitter s,t MOTION BLUR – TEMPORAL JITTERING for subsample get delta time from table jitter delta +/- 1/2 time division move objects to that instant in time DEPTH OF FIELD generate ray from subsample through lens center to focal plane generate random sample on lens disk - random in 2D u,v generate ray from this point to focal plane point VISIBILITY - as usual intersect ray with environment find first intersection at point p on object o with normal n SHADOWS generate random vector on surface of light - random on sphere REFLECTIONS computer reflection vector generate random sample in sphere at end of R TRANSPARENCY compute transmission vector generate random sample in sphere at end of T SIDE NOTE randomize n instead of ramdomize R and T .
    [Show full text]
  • Research on Feature Point Registration Method for Wireless Multi-Exposure Images in Mobile Photography Hui Xu1,2
    Xu EURASIP Journal on Wireless Communications and Networking (2020) 2020:98 https://doi.org/10.1186/s13638-020-01695-4 RESEARCH Open Access Research on feature point registration method for wireless multi-exposure images in mobile photography Hui Xu1,2 Correspondence: [email protected] 1School of Computer Software, Abstract Tianjin University, Tianjin 300072, China In the mobile shooting environment, the multi-exposure is easy to occur due to the 2College of Information impact of the jitter and the sudden change of ambient illumination, so it is necessary Engineering, Henan Institute of to deal with the feature point registration of the multi-exposure image under mobile Science and Technology, Xinxiang 453003, China photography to improve the image quality. A feature point registration technique is proposed based on white balance offset compensation. The global motion estimation of the image is carried out, and the spatial neighborhood information is integrated into the amplitude detection of the multi-exposureimageundermobilephotography,andthe amplitude characteristics of the multi-exposure image under the mobile shooting are extracted. The texture information of the multi-exposure image is compared to that of a global moving RGB 3D bit plane random field, and the white balance deviation of the multi-exposure image is compensated. At different scales, suitable white balance offset compensation function is used to describe the feature points of the multi-exposure image, the parallax analysis and corner detection of the target pixel of the multi-exposure image are carried out, and the image stabilization is realized by combining the feature registration method. The simulation results show that the proposed method has high accuracy and good registration performance for multi-exposure image feature points under mobile photography, and the image quality is improved.
    [Show full text]
  • A Low-Cost Flash Photographic System for Visualization of Droplets
    Journal of Imaging Science and Technology R 62(6): 060502-1–060502-9, 2018. c Society for Imaging Science and Technology 2018 A Low-Cost Flash Photographic System for Visualization of Droplets in Drop-on-Demand Inkjet Huicong Jiang, Brett Andrew Merritt, and Hua Tan School of Engineering and Computer Science, Washington State University Vancouver, 14204 NE Salmon Creek Ave, Vancouver, WA 98686, USA E-mail: [email protected] the extreme, it tries to reduce the flash time of the light source Abstract. Flash photography has been widely used to study which is much easier than improving the shutter speed of the droplet dynamics in drop-on-demand (DoD) inkjet due to its distinct advantages in cost and image quality. However, the a camera and can be compatible with most of the cameras. whole setup, typically comprising the mounting platform, flash Although only one image can be taken from an event through light source, inkjet system, CCD camera, magnification lens and this method, by leveraging the high reproducibility of the pulse generator, still costs tens of thousands of dollars. To reduce DoD inkjet and delaying the flash, a sequence of images the cost of visualization for DoD inkjet droplets, we proposed to replace the expensive professional pulse generator with a freezing different moments can be acquired from a series of low-cost microcontroller board in the flash photographic system. identical events and then yields a video recording the whole The temporal accuracy of the microcontroller was measured by an process. It is noteworthy that flash photography should be oscilloscope.
    [Show full text]
  • Jittered Exposures for Image Super-Resolution
    Jittered Exposures for Image Super-Resolution Nianyi Li1 Scott McCloskey2 Jingyi Yu3,1 1University of Delaware, Newark, DE, USA. [email protected] 2Honeywell ACST, Golden Valley, MN, USA. [email protected] 3ShanghaiTech University, Shanghai, China. [email protected] Abstract The process can be formulated using an observation mod- el that takes low-resolution (LR) images L as the blurred Camera design involves tradeoffs between spatial and result of a high resolution image H: temporal resolution. For instance, traditional cameras pro- vide either high spatial resolution (e.g., DSLRs) or high L = W H + n (1) frame rate, but not both. Our approach exploits the optical stabilization hardware already present in commercial cam- where W = DBM, in which M is the warp matrix (transla- eras and increasingly available in smartphones. Whereas tion, rotation, etc.), B is the blur matrix, D is the decimation single image super-resolution (SR) methods can produce matrix and n is the noise. convincing-looking images and have recently been shown Most state-of-the-art image SR methods consist of three to improve the performance of certain vision tasks, they stages, i.e., registration, interpolation and restoration (an in- are still limited in their ability to fully recover informa- verse procedure). These steps can be implemented separate- tion lost due to under-sampling. In this paper, we present ly or simultaneously according to the reconstruction meth- a new imaging technique that efficiently trades temporal ods adopted. Registration refers to the procedure of estimat- resolution for spatial resolution in excess of the sensor’s ing the motion information or the warp function M between pixel count without attenuating light or adding additional H and L.
    [Show full text]
  • A Novel Technique for Precision Geometric Correction of Jitter Distortion for the Europa Imaging System and Other Rolling-Shutter Cameras
    49th Lunar and Planetary Science Conference 2018 (LPI Contrib. No. 2083) 2188.pdf A NOVEL TECHNIQUE FOR PRECISION GEOMETRIC CORRECTION OF JITTER DISTORTION FOR THE EUROPA IMAGING SYSTEM AND OTHER ROLLING-SHUTTER CAMERAS. M. Shepherd1, R. L. Kirk1 and S. Sides1, 1Astrogeology Science Center, U.S. Geological Survey, 2255 N. Gemini Dr., Flagstaff AZ 86001 ([email protected]). Summary: We use simulated images to demonstrate a Approach: Our approach makes use of the flexibility of novel technique for mitigating geometric distortions caused APS readout. The systematic (line by line) readout of the by platform motion (“jitter”) as two dimensional image sen- frame area is periodically interrupted to read some detector sors are exposed and read out line by line (“rolling shutter”). rows (“check locations”) one or more additional times, re- The results indicate that the Europa Imaging System (EIS) on sulting in “check lines.” These are matched to the corre- NASA’s Europa Clipper can likely meet its scientific goals sponding locations in the systematic readout and the resulting requiring 0.1 pixel precision. The method will also apply to time-differences of pointing are analyzed to produce a model other rolling-shutter cameras. of the pointing history (as in the pushbroom case, the abso- Background: Remote imaging has been a key tool for lute pointing bias cannot be determined from the differ- planetary investigation since the earliest days of the space ences). Because the expected motions are small, it is con- age, but the technology has evolved greatly. Each new type venient to model them as translations in the image plane of sensor (film, vidicon, charge-coupled device or CCD, and rather than as rotations of the camera.
    [Show full text]
  • The Jit.Matrix Object
    JITTER Tutorials and Topics Table of Contents Copyright and Trademark Notices . 9 Credits . 9 About Jitter . .10 Video . .10 2D/3D Graphics . .10 Ease of Use . .10 Matrices . .11 More Details . .11 How to Use The Jitter Documentation . .12 Matrices: What is a Matrix? . 13 A Video Screen is One Type of Matrix . .14 What is a Plane? . .16 The Data in a Matrix . .16 Attributes: Editing Jitter object parameters . 19 What are Attributes? . .19 Setting Attributes . .20 Jitter Object Arguments . .21 Querying Attributes and Object State . .22 Summary . .24 Tutorial 1: Playing a QuickTime Movie . 25 How Jitter Objects Communicate . .26 Causing Action by Jitter Objects . .26 Arguments in the Objects . .28 Summary . .29 Tutorial 2: Create a Matrix . 30 What's a Matrix? . .30 The jit.matrix object . .30 The jit.print Object . .31 Setting and Querying Values in a Matrix . .32 The jit.pwindow Object — . .34 Filling a Matrix Algorithmically . .35 Other jit.matrix Messages . .36 Summary . .36 Tutorial 3: Math Operations on a Matrix . 38 Operation @-Sign . .39 Math Operations on Multiple Planes of Data . .40 Modifying the Colors in an Image . .41 Sizing it Up . .43 Summary . .44 1 Table of Contents Tutorial 4: Controlling Movie Playback . 46 Obtaining Some Information About the Movie . .47 Starting, Stopping, and Slowing Down . .48 Time is on My Side . .49 Scrubbing and Looping . .50 Summary . .50 Tutorial 5: ARGB Color . 52 Color in Jitter . .52 Color Components: RGB . .52 The Alpha Channel . .52 Color Data: char, long, or float . .53 Isolating Planes of a Matrix . .53 Color Rotation .
    [Show full text]
  • Distribution Ray Tracing Academic Press, 1989
    Reading Required: Shirley, 13.11, 14.1-14.3 Further reading: A. Glassner. An Introduction to Ray Tracing. Distribution Ray Tracing Academic Press, 1989. [In the lab.] Robert L. Cook, Thomas Porter, Loren Carpenter. “Distributed Ray Tracing.” Computer Graphics Brian Curless (Proceedings of SIGGRAPH 84). 18 (3). pp. 137- CSE 557 145. 1984. Fall 2015 James T. Kajiya. “The Rendering Equation.” Computer Graphics (Proceedings of SIGGRAPH 86). 20 (4). pp. 143-150. 1986. 1 2 Simulating gloss and translucency Soft shadows The mirror-like form of reflection, when used to approximate glossy surfaces, introduces a kind of aliasing, because we are under-sampling reflection (and refraction). For example: Distributing rays over light source area gives: Distributing rays over reflection directions gives: 3 4 The pinhole camera Shrinking the pinhole The first camera - “camera obscura” - known to Aristotle. In 3D, we can visualize the blur induced by the pinhole (a.k.a., aperture): Q: What happens as we continue to shrink the aperture? Q: How would we reduce blur? 5 6 Shrinking the pinhole, cont’d The pinhole camera, revisited We can think in terms of light heading toward the image plane: We can equivalently turn this around by following rays from the viewer: 7 8 The pinhole camera, revisited Lenses Pinhole cameras in the real world require small Given this flipped version: apertures to keep the image in focus. Lenses focus a bundle of rays to one point => can have larger aperture. how can we simulate a pinhole camera more accurately? For a “thin” lens, we can approximately calculate where an object point will be in focus using the the Gaussian lens formula: 111 ddio f where f is the focal length of the lens.
    [Show full text]
  • Software Simulation of Depth of Field Effects in Video from Small Aperture Cameras Jordan Sorensen
    Software Simulation of Depth of Field Effects in Video from Small Aperture Cameras MASSACHUSETTS INSTITUTE by OF TECHNOLOGY Jordan Sorensen AUG 2 4 2010 B.S., Electrical Engineering, M.I.T., 2009 LIBRARIES Submitted to the Department of Electrical Engineering and Computer Science in partial fulfillment of the requirements for the degree of ARCHNES Master of Engineering in Electrical Engineering and Computer Science at the MASSACHUSETTS INSTITUTE OF TECHNOLOGY June 2010 @ Massachusetts Institute of Technology 2010. All rights reserved. Author .. Depi4nfent of Electrical Engineering and Computer Science May 21, 2010 Certified by....................................... ... Ramesh Raskar Associate Professor, MIT Media Laboratory Thesis Supervisor Accepted by. ...... ... ... ..... .... ..... .A .. Dr. Christopher YOerman Chairman, Department Committee on Graduate Theses 2 Software Simulation of Depth of Field Effects in Video from Small Aperture Cameras by Jordan Sorensen Submitted to the Department of Electrical Engineering and Computer Science on May 21, 2010, in partial fulfillment of the requirements for the degree of Master of Engineering in Electrical Engineering and Computer Science Abstract This thesis proposes a technique for post processing digital video to introduce a simulated depth of field effect. Because the technique is implemented in software, it affords the user greater control over the parameters of the effect (such as the amount of defocus, aperture shape, and defocus plane) and allows the effect to be used even on hardware which would not typically allow for depth of field. In addition, because it is a completely post processing technique and requires no change in capture method or hardware, it can be used on any video and introduces no new costs.
    [Show full text]
  • Noordegraaf DEF.Indd 1 08-01-13 14:33 FRAMING FILM
    FRAMING ET AL. (EDS.) NOORDEGRAAF JULIA NOORDEGRAAF, COSETTA G. SABA, FILM BARBARA LE MAÎTRE, VINZENZ HEDIGER (EDS.) PRESERVING AND EXHIBITING MEDIA ART PRESERVING AND EXHIBITING MEDIA ART Challenges and Perspectives This important and fi rst-of-its-kind collection addresses the Julia Noordegraaf is AND EXHIBITING MEDIA ART PRESERVING emerging challenges in the fi eld of media art preservation and Professor of Heritage and exhibition, providing an outline for the training of professionals in Digital Culture at the Univer- this fi eld. Since the emergence of time-based media such as fi lm, sity of Amsterdam. video and digital technology, artists have used them to experiment Cosetta G. Saba is Associate with their potential. The resulting artworks, with their basis in Professor of Film Analysis rapidly developing technologies that cross over into other domains and Audiovisual Practices in such as broadcasting and social media, have challenged the tradi- Media Art at the University tional infrastructures for the collection, preservation and exhibition of Udine. of art. Addressing these challenges, the authors provide a historical Barbara Le Maître is and theoretical survey of the fi eld, and introduce students to the Associate Professor of challenges and di culties of preserving and exhibiting media art Theory and Aesthetics of through a series of fi rst-hand case studies. Situated at the threshold Static and Time-based Images at the Université Sorbonne between archival practices and fi lm and media theory, it also makes nouvelle – Paris 3. a strong contribution to the growing literature on archive theory Vinzenz Hediger is and archival practices. Professor of Film Studies at the Goethe-Universität Frankfurt am Main.
    [Show full text]
  • Coverage-Based Opacity Estimation for Interactive Depth of Field in Molecular Visualization
    Coverage-Based Opacity Estimation for Interactive Depth of Field in Molecular Visualization Sathish Kottravel∗ Martin Falk∗ Erik Sunden´ ∗ Timo Ropinski† ∗ Interactive Visualization Group, Linkoping¨ University, Sweden † Visual Computing Research Group, Ulm University, Germany Figure 1: Thermus thermophilus 70S ribosome (PDB ID: 2WDK) rendered with no Depth of Field effect (left) and using our approach focusing on near structures (center) and far away structures (right). ABSTRACT 1 INTRODUCTION To better understand the underlying mechanisms of life, it is es- In this paper, we introduce coverage-based opacity estimation to sential to investigate the structure of molecules. Considering the achieve Depth of Field (DoF) effects when visualizing molec- structure-follows-function paradigm, the understanding of a pro- ular dynamics (MD) data. The proposed algorithm is a novel tein’s structure for instance can give essential hints about its role object-based approach which eliminates many of the shortcom- in metabolic pathways. Furthermore, when dealing with more than ings of state-of-the-art image-based DoF algorithms. Based on one molecule, their structure can give us clues about potential bind- observations derived from a physically-correct reference renderer, ing sites for molecule interactions. Besides those cases requir- coverage-based opacity estimation exploits semi-transparency to ing depictions of individual molecules, visualizing the multitude simulate the blur inherent to DoF effects. It achieves high qual- of molecules within the crowded environment of the cell also calls ity DoF effects, by augmenting each atom with a semi-transparent for improved spatial comprehension. While modern life science shell, which has a radius proportional to the distance from the fo- technologies result in detailed molecular models as well as sim- cal plane of the camera.
    [Show full text]