Improving Resolution and Depth-Of-Field of Light Field Cameras Using a Hybrid Imaging System
Total Page:16
File Type:pdf, Size:1020Kb
Improving Resolution and Depth-of-Field of Light Field Cameras Using a Hybrid Imaging System Vivek Boominathan, Kaushik Mitra, Ashok Veeraraghavan Rice University 6100 Main St, Houston, TX 77005 [vivekb, Kaushik.Mitra, vashok] @rice.edu Abstract Lytro Hybrid Imager camera (our system) Current light field (LF) cameras provide low spatial res- olution and limited depth-of-field (DOF) control when com- 11 megapixels 0.1 megapixels 9 X 9 angular pared to traditional digital SLR (DSLR) cameras. We show 9 X 9 angular resolution. that a hybrid imaging system consisting of a standard LF resolution. camera and a high-resolution (HR) standard camera en- ables (a) achieve high-resolution digital refocusing, (b) bet- Fundamental Angular Angular Resolution resolution trade-off ter DOF control than LF cameras, and (c) render grace- 11 megapixels DSLR no angular ful high-resolution viewpoint variations, all of which were camera previously unachievable. We propose a simple patch-based information. algorithm to super-resolve the low-resolution (LR) views of Spatial Resolution the light field using the high-resolution patches captured us- Figure 1: Fundamental resolution trade-off in light-field ing a HR SLR camera. The algorithm does not require the imaging: Given a fixed resolution sensor there is an inverse LF camera and the DSLR to be co-located or for any cali- relationship between spatial resolution and angular resolu- bration information regarding the two imaging systems. We tion that can be captured. By using a hybrid imaging system build an example prototype using a Lytro camera (380×380 containing two sensors, one a high spatial resolution camera pixel spatial resolution) and a 18 megapixel (MP) Canon and another a light-field camera, one can reconstruct a high DSLR camera to generate a light field with 11 MP reso- resolution light field. 1 th lution (9× super-resolution) and about 9 of the DOF of the Lytro camera. We show several experimental results on 1.1. Motivation challenging scenes containing occlusions, specularities and Fundamental Resolution Trade-off: Given an image complex non-lambertian materials, demonstrating the effec- sensor of a fixed resolution, existing methods for captur- tiveness of our approach. ing light-field trade-off spatial resolution in order to acquire angular resolution (with the exception of the recently pro- 1. Introduction posed mask-based method [25]). Consider an example of a Light field (LF) is a 4-D function that measures the spa- 11 MP image sensor, much like the one used in the Lytro tial and angular variations in the intensity of light [2]. Ac- camera [23]. A traditional camera using such a sensor is quiring light fields provides us with three important capa- capable of recording 11 MP images, but acquires no an- bilities that a traditional camera does not allow: (1) ren- gular information and therefore provides no ability to per- der images with small viewpoint changes, (2) render im- form post-capture refocusing. In contrast, the Lytro camera ages with post-capture control of focus and depth-of-field, is capable of recording 9 × 9 angular resolution, but has and (3) compute a depth map or a range image by utilizing a low spatial resolution of 380 × 380 pixels (since it con- either multi-view stereo or depth from focus/defocus meth- tains a 9 × 9 pixels per lens array). This resolution loss is ods. The growing popularity of LF cameras is attributed not restricted to microlens array-based LF cameras but is a to these three novel capabilities. Nevertheless, current LF common handicap faced by other LF cameras including tra- cameras suffer from two significant limitations that hamper ditional mask-based LF cameras [34], camera arrays [29], their widespread appeal and adoption: (1) low spatial reso- and angle-sensitive pixels [35]. Thus, there is an imminent lution and (2) limited DOF control. need for improving the spatial resolution characteristics of Near Focused Far Focused Depth Map DSLR Camera (high-resolution image) Light-field Camera (low-resolution light-field) + Hybrid Imaging (high-resolution light-field) Depth Map Near Focused Far Focused Figure 2: Traditional cameras capture high resolution photographs but provide no post-capture focusing controls. Common light field cameras provide depth maps and post-capture refocusing ability but at very low spatial resolution. Here, we show that a hybrid imaging system comprising of a high-resolution camera and a light-field camera allows us to obtain high-resolution depth maps and post-capture refocusing. LF sensors. We propose using a hybrid imaging system con- 1.2. Contributions taining two cameras, one with a high spatial resolution sen- In this paper, we propose a hybrid imaging system con- sor and the second being a light-field camera, which can be sisting of a high resolution standard camera along with the used to reconstruct a high resolution light field (see Figure low-resolution LF camera. This hybrid imaging system 1). (Figure 2) along with the associated algorithms enables us to capture/render (a) high spatial resolution light field, (b) Depth Map Resolution: LF cameras enable to compu- high spatial resolution depth maps, (c) higher depth resolu- tation of depth information by the application of multi-view tion (more depth layers), and (d) shallower DOF. stereo or depth from focus/defocus methods on the rendered views. Unfortunately, the low spatial resolution of the ren- 2. Related Work dered views result in low resolution depth maps. In addi- LF capture: Existing LF cameras can be divided into tion, since the depth resolution of the depth maps (i.e., the two main categories: (a) single shot [28, 18, 34, 23, 30, 17, number of distinct depth profiles within a fixed imaging vol- 29, 35], and (b) multiple shot [21, 4]. Single shot light field ume) is directly proportional to the disparity between views, cameras multiplex the 4-D LF onto the 2D sensor, losing the low resolution of the views directly result in very few spatial resolution to capture the angular information in the depth layers in the recovered range map. This results in er- LF. Such cameras employ either a lenslet array close to the rors when the depth information is directly used for vision sensor [28, 17], a mask close to the sensor [34], angle sen- tasks such as segmentation and object/activity recognition. sitive pixels [35] or an array of lens/prism outside the main lens [18]. An example of multiple shot LF capture is pro- DOF Control: The DOF of an imaging system is in- grammable aperture imaging [21], which allows capturing versely proportional to the image resolution (for a fixed f# light fields at the spatial resolution of the sensor. Recently, and sensor size). Since the rendered views in a LF cam- Babacan et al. [4], Marwah et al. [25] and Tambe et al. [32] era are low-resolution, this results in much larger DOF than show that one can use compressive sensing and dictionary can be attained using high resolution DSLR cameras with learning to reduce the number of images required. The rein- similar sensor size and f#. This is a primary reason why terpretable imager by Agrawal et al. [3] has shown resolu- DSLR cameras provide shallower DOF and are favored for tion trade-offs in a single image capture. Another approach photography. for capturing LF is to use a camera array [19, 20, 37]. How- ever, such approaches are hardware intensive, costly and re- There is a need for a high resolution shallow DOF LF quire extensive bandwidth, storage and power consumption. imaging device. LF Super-resolution and Plenoptic2:0: The Plenoptic2:0 camera [17] recovers the lost resolution the self-similarity within the input image [15]. In our hy- by placing the microlens array at a different location brid imaging method, patches from each view of a LF is compared to the original design [28]. Similarly, the matched with a reference high resolution image the same Raytrix camera [30] uses a microlens array with lenses scene. Since the high resolution image has the exact de- of different focal length to improve spatial resolution. tails of the scene, the super-resolved LF has the true infor- Recently, several LF super-resolution algorithms have mation compared to hallucinated information by [16, 15]. been proposed to recover the lost resolution [7, 36, 26]. Recently, a couple of fast approximate nearest patch search Apart from these hardware modifications to the plenoptic algorithms have been introduced [5]. We use the fast library camera, super-resolution algorithms in context of LF have for approximate nearest neighbors (FLANN) [27] to search also been proposed. Bishop et al. [7] proposed a Bayesian for matching patches in the reference high-resolution im- framework in which they assume Lambertian textural age. priors in the image formation model and estimate both the high resolution depth map and light field. Wanner et al. 3. Hybrid Light Field Imaging [36] propose to compute continuous disparity maps using The hybrid imager we propose is a combination of two the epipolar plane image (EPI) structure of the LF. They imaging systems: a low resolution LF device (Lytro cam- then use this disparity map and variational techniques to era) and a high-resolution camera (DSLR). The Lytro cam- compute super-resolved novel views. Mitra et al. [26] learn era captures the angular perspective views and the depth of Gaussian mixture model (GMM) for light field patches and the scene while the DSLR camera captures a photograph of perform Bayesian inference to obtain super-resolved LF. the scene. Our algorithm combines these two imaging sys- Most of these methods show modest super-resolution by tems to produce a light field with the spatial resolution of a factor of 4×.