Rendering Natural Camera Bokeh Effect with Deep Learning

Total Page:16

File Type:pdf, Size:1020Kb

Rendering Natural Camera Bokeh Effect with Deep Learning Rendering Natural Camera Bokeh Effect with Deep Learning Andrey Ignatov Jagruti Patel Radu Timofte [email protected] [email protected] [email protected] ETH Zurich, Switzerland Abstract Bokeh is an important artistic effect used to highlight the main object of interest on the photo by blurring all out-of-focus areas. While DSLR and system camera lenses can render this effect naturally, mobile cameras are unable to produce shallow depth-of-field photos due to a very small aperture diameter of their optics. Unlike the current solutions simulating bokeh by applying Gaussian blur to image background, in this paper we propose to learn a realistic shallow focus technique directly from the photos produced by DSLR cameras. For this, we present a large-scale bokeh dataset consisting of 5K shallow / wide depth-of-field image pairs captured using the Canon 7D DSLR with 50mm f/1.8 lenses. We use these images to train a deep learning model to reproduce a natural bokeh effect based on a single narrow-aperture image. The experimental results show that the proposed approach is able to render a plausible non-uniform bokeh even in case of complex input data with multiple objects. The dataset, pre-trained models and codes used in this paper are available on the project website: https://people.ee.ethz.ch/˜ihnatova/ pynet-bokeh.html 1. Introduction Figure 1. The original shallow depth-of-field image and the image produced with our method. Bokeh effect is a very popular photography technique used to make the subject in the shot stand out sharply unattainable for mobile cameras with compact optics and against a blurred background (Fig.1). It is achieved by fo- tiny sensors. As a result, bokeh effect can only be sim- cusing the camera on the selected area or object and shoot- ulated computationally on smartphones and other devices ing the photo with a wide aperture lens. This produces with small mobile cameras. a shallow depth-of-field image where only objects located Synthetic bokeh effect rendering is a relatively new ma- within a narrow image plane are visible clearly, while all chine learning topic that has emerged in the past years. other parts of the image are blurred. This effect is often [15, 16] were one of the first papers exploring shallow leading to very pleasing visual results, and besides that al- depth-of-field simulation based on a single image. These lows to wash out unnecessary, distracting or unattractive works were focused on portrait photos only: they first seg- background details, which is especially useful in case of mented out people from the image with a convolutional neu- mobile photography. In order to get good bokeh, fast lenses ral network, and then uniformly blurred the remaining im- with a large aperture are needed, which makes this effect age background. The same approach was also considered 1 Figure 2. Sample wide and shallow depth-of-field image pairs from the EBB! dataset. in [21], where the authors proposed an efficient solution for In this paper, we present a different approach to the con- person segmentation. There also exists a separate group of sidered problem. Since the properties of the real bokeh de- works [8,2, 19] that are blurring the image based on the pend on a large number of parameters such as focal length, predicted depth map, however they require the camera to distance to the subject, the acceptable circle of confusion, move in order to get the parallax effect, thus the practical the aperture size, type and construction, various optical lens applicability of these approaches is quite limited. An accu- characteristics, etc., it is almost impossible to render this rate depth map estimation can be also obtained with stereo effect precisely with the existing software tools and meth- vision [1] using synchronously captured photos from multi- ods. Therefore, in this work we propose to learn the bokeh ple cameras. technique directly from the photos produced by a high-end Since 2015, the majority of flagship and mid-range mo- DSLR camera using deep learning. The presented approach bile devices are getting special hardware for accelerating is camera independent, does not require any special hard- deep learning models [5]. The performance of such hard- ware, and can also be applied to the existing images. ware is already approaching the results of mid-range Nvidia GPUs presented in the past years [6], making the use of Our main contributions are: deep learning-based methods for bokeh effect simulation on • A novel end-to-end deep learning solution for realis- mobile devices especially relevant. Paper [17] describes the tic bokeh effect rendering. The model is trained to synthetic depth-of-field rendering approach used on Google map the standard narrow-aperture images into shallow Pixel devices that is able to achieve very impressive results depth-of-field photos captured with a DSLR camera. for the majority of photo scenes. In this work, the authors are predicting a depth map based on a single image and are • A large-scale dataset containing 5K shallow / wide using camera’s dual-pixel auto-focus system when possible depth-of-field image pairs collected in the wild with for refinement of the results. The weakest part of the paper the Canon 7D DSLR camera and 50mm f/1.8 fast lens. is the actual bokeh rendering method – a simple approxi- • A comprehensive set of experiments evaluating the mated disk blur is used for this, making the visual results quantitative and perceptual quality of the rendered quite different from the natural bokeh photos produced by bokeh photos. We demonstrate that the proposed solu- DSLR cameras. tion can process one 1024×1536 px image under 5s on all high-end mobile chipsets when running it on GPU. 2. Everything is Better with Bokeh! Dataset One of the biggest challenges in the bokeh rendering task is to get high-quality real data that can be used for training deep models. To tackle this problem, a large-scale Every- thing is Better with Bokeh! (EBB!) dataset containing more than 10 thousand images was collected in the wild during several months. By controlling the aperture size of the lens, images with shallow and wide depth-of-field were taken. In each photo pair, the first image was captured with a narrow aperture (f/16) that results in a normal sharp photo, whereas the second one was shot using the highest aperture (f/1.8) leading to a strong bokeh effect. The photos were taken during the daytime in a wide variety of places and in vari- ous illumination and weather conditions. The photos were captured in automatic mode, the default settings were used throughout the entire collection procedure. An example set of collected images is presented in Figure2. The captured image pairs are not aligned exactly, there- fore they were first matched using SIFT keypoints and RANSAC method same as in [3]. The resulting images were then cropped to their intersection part and downscaled so that their final height is equal to 1024 pixels. Finally, we computed a coarse depth map for each wide depth-of-field image using the Megadepth model proposed in [12]. These maps can be stacked directly with the input images and used as an additional guidance for the trained model. From Figure 3. The architecture of the PyNET-based model. Concat and the resulting 10 thousand images, 200 image pairs were re- Sum ops are applied to the outputs of the adjacent layers. served for testing, while the other 4.8 thousand photo pairs can be used for training and validation. layers are then concatenated, which allows the network to 3. Proposed Method learn a more diverse set of features at each level. The out- puts obtained at lower scales are upsampled with transposed Bokeh effect simulation problem belongs to a group of convolutional layers, stacked with feature maps from the tasks dealing with both global and local image processing. upper level and then subsequently processed in the follow- High-level image analysis is needed here to detect the ar- ing convolutional layers. Leaky ReLU activation function eas on the photo where the bokeh effect should be applied, is applied after each convolutional operation, except for the whereas low-level processing is used for rendering the ac- output layers that are using tanh function to map the results tual shallow depth-of-field images and refining the results. to (-1, 1) interval. Instance normalization is used in all con- Therefore, in this work we base our solution on the PyNET volutional layers that are processing images at lower scales architecture [7] designed specifically for this kind of tasks: (levels 2-5). We are additionally using two transposed con- it is processing the image at different scales and combining volutional layers on top of the main model that upsample the learned global and local features together. the images to their target size. The model is trained sequentially, starting from the low- 3.1. PyNET CNN Architecture est layer. This allows to achieve good semantically-driven Figure3 illustrates schematic representation of the reconstruction results at smaller scales that are working PyNET-based architecture used in this work. The model has with images of very low resolution and thus performing an inverted pyramidal shape and is processing the images mostly global image manipulations. After the bottom layer at seven different scales. The proposed architecture has a is pre-trained, the same procedure is applied to the next level number of blocks that are processing feature maps in paral- till the training is done on the original resolution.
Recommended publications
  • Depth-Aware Blending of Smoothed Images for Bokeh Effect Generation
    1 Depth-aware Blending of Smoothed Images for Bokeh Effect Generation Saikat Duttaa,∗∗ aIndian Institute of Technology Madras, Chennai, PIN-600036, India ABSTRACT Bokeh effect is used in photography to capture images where the closer objects look sharp and every- thing else stays out-of-focus. Bokeh photos are generally captured using Single Lens Reflex cameras using shallow depth-of-field. Most of the modern smartphones can take bokeh images by leveraging dual rear cameras or a good auto-focus hardware. However, for smartphones with single-rear camera without a good auto-focus hardware, we have to rely on software to generate bokeh images. This kind of system is also useful to generate bokeh effect in already captured images. In this paper, an end-to-end deep learning framework is proposed to generate high-quality bokeh effect from images. The original image and different versions of smoothed images are blended to generate Bokeh effect with the help of a monocular depth estimation network. The proposed approach is compared against a saliency detection based baseline and a number of approaches proposed in AIM 2019 Challenge on Bokeh Effect Synthesis. Extensive experiments are shown in order to understand different parts of the proposed algorithm. The network is lightweight and can process an HD image in 0.03 seconds. This approach ranked second in AIM 2019 Bokeh effect challenge-Perceptual Track. 1. Introduction tant problem in Computer Vision and has gained attention re- cently. Most of the existing approaches(Shen et al., 2016; Wad- Depth-of-field effect or Bokeh effect is often used in photog- hwa et al., 2018; Xu et al., 2018) work on human portraits by raphy to generate aesthetic pictures.
    [Show full text]
  • Portraiture, Surveillance, and the Continuity Aesthetic of Blur
    Michigan Technological University Digital Commons @ Michigan Tech Michigan Tech Publications 6-22-2021 Portraiture, Surveillance, and the Continuity Aesthetic of Blur Stefka Hristova Michigan Technological University, [email protected] Follow this and additional works at: https://digitalcommons.mtu.edu/michigantech-p Part of the Arts and Humanities Commons Recommended Citation Hristova, S. (2021). Portraiture, Surveillance, and the Continuity Aesthetic of Blur. Frames Cinema Journal, 18, 59-98. http://doi.org/10.15664/fcj.v18i1.2249 Retrieved from: https://digitalcommons.mtu.edu/michigantech-p/15062 Follow this and additional works at: https://digitalcommons.mtu.edu/michigantech-p Part of the Arts and Humanities Commons Portraiture, Surveillance, and the Continuity Aesthetic of Blur Stefka Hristova DOI:10.15664/fcj.v18i1.2249 Frames Cinema Journal ISSN 2053–8812 Issue 18 (Jun 2021) http://www.framescinemajournal.com Frames Cinema Journal, Issue 18 (June 2021) Portraiture, Surveillance, and the Continuity Aesthetic of Blur Stefka Hristova Introduction With the increasing transformation of photography away from a camera-based analogue image-making process into a computerised set of procedures, the ontology of the photographic image has been challenged. Portraits in particular have become reconfigured into what Mark B. Hansen has called “digital facial images” and Mitra Azar has subsequently reworked into “algorithmic facial images.” 1 This transition has amplified the role of portraiture as a representational device, as a node in a network
    [Show full text]
  • Using Depth Mapping to Realize Bokeh Effect with a Single Camera Android Device EE368 Project Report Authors (SCPD Students): Jie Gong, Ran Liu, Pradeep Vukkadala
    Using Depth Mapping to realize Bokeh effect with a single camera Android device EE368 Project Report Authors (SCPD students): Jie Gong, Ran Liu, Pradeep Vukkadala Abstract- In this paper we seek to produce a bokeh Bokeh effect is usually achieved in high end SLR effect with a single image taken from an Android device cameras using portrait lenses that are relatively large in size by post processing. Depth mapping is the core of Bokeh and have a shallow depth of field. It is extremely difficult effect production. A depth map is an estimate of depth to achieve the same effect (physically) in smart phones at each pixel in the photo which can be used to identify which have miniaturized camera lenses and sensors. portions of the image that are far away and belong to However, the latest iPhone 7 has a portrait mode which can the background and therefore apply a digital blur to the produce Bokeh effect thanks to the dual cameras background. We present algorithms to determine the configuration. To compete with iPhone 7, Google recently defocus map from a single input image. We obtain a also announced that the latest Google Pixel Phone can take sparse defocus map by calculating the ratio of gradients photos with Bokeh effect, which would be achieved by from original image and reblured image. Then, full taking 2 photos at different depths to camera and defocus map is obtained by propagating values from combining then via software. There is a gap that neither of edges to entire image by using nearest neighbor method two biggest players can achieve Bokeh effect only using a and matting Laplacian.
    [Show full text]
  • AG-AF100 28Mm Wide Lens
    Contents 1. What change when you use the different imager size camera? 1. What happens? 2. Focal Length 2. Iris (F Stop) 3. Flange Back Adjustment 2. Why Bokeh occurs? 1. F Stop 2. Circle of confusion diameter limit 3. Airy Disc 4. Bokeh by Diffraction 5. 1/3” lens Response (Example) 6. What does In/Out of Focus mean? 7. Depth of Field 8. How to use Bokeh to shoot impressive pictures. 9. Note for AF100 shooting 3. Crop Factor 1. How to use Crop Factor 2. Foal Length and Depth of Field by Imager Size 3. What is the benefit of large sensor? 4. Appendix 1. Size of Imagers 2. Color Separation Filter 3. Sensitivity Comparison 4. ASA Sensitivity 5. Depth of Field Comparison by Imager Size 6. F Stop to get the same Depth of Field 7. Back Focus and Flange Back (Flange Focal Distance) 8. Distance Error by Flange Back Error 9. View Angle Formula 10. Conceptual Schema – Relationship between Iris and Resolution 11. What’s the difference between Video Camera Lens and Still Camera Lens 12. Depth of Field Formula 1.What changes when you use the different imager size camera? 1. Focal Length changes 58mm + + It becomes 35mm Full Frame Standard Lens (CANON, NIKON, LEICA etc.) AG-AF100 28mm Wide Lens 2. Iris (F Stop) changes *distance to object:2m Depth of Field changes *Iris:F4 2m 0m F4 F2 X X <35mm Still Camera> 0.26m 0.2m 0.4m 0.26m 0.2m F4 <4/3 inch> X 0.9m X F2 0.6m 0.4m 0.26m 0.2m Depth of Field 3.
    [Show full text]
  • Intro to Digital Photography.Pdf
    ABSTRACT Learn and master the basic features of your camera to gain better control of your photos. Individualized chapters on each of the cameras basic functions as well as cheat sheets you can download and print for use while shooting. Neuberger, Lawrence INTRO TO DGMA 3303 Digital Photography DIGITAL PHOTOGRAPHY Mastering the Basics Table of Contents Camera Controls ............................................................................................................................. 7 Camera Controls ......................................................................................................................... 7 Image Sensor .............................................................................................................................. 8 Camera Lens .............................................................................................................................. 8 Camera Modes ............................................................................................................................ 9 Built-in Flash ............................................................................................................................. 11 Viewing System ........................................................................................................................ 11 Image File Formats ....................................................................................................................... 13 File Compression ......................................................................................................................
    [Show full text]
  • The Daguerreotype Achromat 2.9/64 Art Lens by Lomography Is a Revival of This Lost Aesthetic
    OFFICIAL PRESS RELEASE The Lomographic Society Proudly Presents: Press Release Date: 6th April 2016 The DAGUERREOTYPE PRESS CONTACT Katherine Phipps [email protected] ACHROMAT 2.9/64 ART LENS Lomographic Society International 41 W 8th St The Great Return of a Lost Aesthetic from 1839 New York, NY 10011 - for Modern Day Cameras LINKS FOR EDITORS Back us on Kickstarter Now: www.lomography.com/kickstarter · The Ethereal Aesthetics of The World’s First Visit the Daguerreotype Achromat Photographic Optic Lens: reviving an optical design 2.9/64 Art Lens Microsite: www.lomography.com/achromat from 1839 by Daguerre and Chevalier · A Highly Versatile Tool for Modern Day Cameras: offering crisp sharpness, silky soft focus, freedom to control depth of field with multiple bokeh effects · Premium Quality Craftsmanship of The Art Lens Family: designed and handcrafted in a small manufactory, available in black and brass finish – reinvented for use with modern digital and analogue cameras · Proud to be Back on Kickstarter: following Lomography’s history of successful campaigns and exciting projects The Ethereal Aesthetics Of The World’s First Optic Lens Practical photography was invented in 1839, with the combination of a Chevalier Achromat Lens attached to a Daguerreotype camera. The signature character of the Chevalier lens bathed images in an alluring veil of light, due to a series of beautiful “aberrations” in its image-forming optical system, which naturally caused a glazy, soft picture at wide apertures. The Daguerreotype Achromat 2.9/64 Art Lens by Lomography is a revival of this lost aesthetic. Today, more photographs are taken every two minutes than humanity took in the entire 19th century.
    [Show full text]
  • Three Techniques for Rendering Generalized Depth of Field Effects
    Three Techniques for Rendering Generalized Depth of Field Effects Todd J. Kosloff∗ Computer Science Division, University of California, Berkeley, CA 94720 Brian A. Barskyy Computer Science Division and School of Optometry, University of California, Berkeley, CA 94720 Abstract Post-process methods are fast, sometimes to the point Depth of field refers to the swath that is imaged in sufficient of real-time [13, 17, 9], but generally do not share the focus through an optics system, such as a camera lens. same image quality as distributed ray tracing. Control over depth of field is an important artistic tool that can be used to emphasize the subject of a photograph. In a A full literature review of depth of field methods real camera, the control over depth of field is limited by the is beyond the scope of this paper, but the interested laws of physics and by physical constraints. Depth of field reader should consult the following surveys: [1, 2, 5]. has been rendered in computer graphics, but usually with the same limited control as found in real camera lenses. In Kosara [8] introduced the notion of semantic depth this paper, we generalize depth of field in computer graphics of field, a somewhat similar notion to generalized depth by allowing the user to specify the distribution of blur of field. Semantic depth of field is non-photorealistic throughout a scene in a more flexible manner. Generalized depth of field provides a novel tool to emphasize an area of depth of field used for visualization purposes. Semantic interest within a 3D scene, to select objects from a crowd, depth of field operates at a per-object granularity, and to render a busy, complex picture more understandable allowing each object to have a different amount of blur.
    [Show full text]
  • Photography Techniques Intermediate Skills
    Photography Techniques Intermediate Skills PDF generated using the open source mwlib toolkit. See http://code.pediapress.com/ for more information. PDF generated at: Wed, 21 Aug 2013 16:20:56 UTC Contents Articles Bokeh 1 Macro photography 5 Fill flash 12 Light painting 12 Panning (camera) 15 Star trail 17 Time-lapse photography 19 Panoramic photography 27 Cross processing 33 Tilted plane focus 34 Harris shutter 37 References Article Sources and Contributors 38 Image Sources, Licenses and Contributors 39 Article Licenses License 41 Bokeh 1 Bokeh In photography, bokeh (Originally /ˈboʊkɛ/,[1] /ˈboʊkeɪ/ BOH-kay — [] also sometimes heard as /ˈboʊkə/ BOH-kə, Japanese: [boke]) is the blur,[2][3] or the aesthetic quality of the blur,[][4][5] in out-of-focus areas of an image. Bokeh has been defined as "the way the lens renders out-of-focus points of light".[6] However, differences in lens aberrations and aperture shape cause some lens designs to blur the image in a way that is pleasing to the eye, while others produce blurring that is unpleasant or distracting—"good" and "bad" bokeh, respectively.[2] Bokeh occurs for parts of the scene that lie outside the Coarse bokeh on a photo shot with an 85 mm lens and 70 mm entrance pupil diameter, which depth of field. Photographers sometimes deliberately use a shallow corresponds to f/1.2 focus technique to create images with prominent out-of-focus regions. Bokeh is often most visible around small background highlights, such as specular reflections and light sources, which is why it is often associated with such areas.[2] However, bokeh is not limited to highlights; blur occurs in all out-of-focus regions of the image.
    [Show full text]
  • Backlighting
    Challenge #1 - Capture Light Bokeh Have you seen those beautiful little circles of colour and light in the background of photos? It’s called Bokeh and it comes from the Japanese word “boke” meaning blur. It’s so pretty and not only does it make a gorgeous backdrop for portraits, but it can be the subject in it’s own right too! It’s also one of the most fun aspects of learning photography, being able to capture your own lovely bokeh! And that’s what our first challenge is all about, Capturing Light Bokeh! And remember - just as I said in this video, the purpose of these challenges is NOT to take perfect photos the first time you try something new… it’s about looking for light, trying new techniques, and exploring your creativity to help you build your skills and motivate and inspire you. So lets have fun & I can’t wait to see your photos! Copyright 2017 - www.clicklovegrow.com How to Achieve Light Bokeh Light Bokeh is created when light reflects off, or through, a background; and is then captured with a wide open aperture of your lens. As light reflects differently off flat surfaces, you’re not likely to see this effect when using a plain wall, for example. Instead, you’ll see it when your background has a little texture, such as light reflecting off leaves in foliage of a garden, or wet grass… or when light is broken up when streaming through trees. Looking for Light One of the key take-aways for this challenge is to start paying attention to light & how you can capture it in photos… both to add interest and feeling to your portraits, but as a creative subject in it’s own right.
    [Show full text]
  • Introduction to Aperture in Digital Photography Helpful
    9/7/2015 Introduction to Aperture - Digital Photography School Tips & Tutorials Cameras & Equipment Post Production Books Presets Courses Forums Search Introduction to Aperture in Join 1,417,294 Subscribers! Digital Photography A Post By: Darren Rowse 142K 23.1K 661 164 1200 96 0 35 Shares Share Share Share Pin it Share Other Comments Photography Tips & Tutorials Tips for Getting Started with Urban Landscape Photography 6 Tips for Creating More Captivating Landscape Photographs How to Solve 5 Composition Tempest by Sheldon Evans on 500px Conundrums Faced by Over the last couple of weeks I’ve been writing a series of posts on elements Landscape Photographers that digital photographers need to learn about in order to get out of Auto mode and learn how to manually set the exposure of their shots. I’ve largely focussed Landscape Photography - upon three elements of the ‘exposure triangle‘ – ISO, Shutter Speed and Shooting the Same Aperture. I’ve previously written about the first two and today would like to Location Through the turn our attention to Aperture. Seasons Before I start with the explanations let me say this. If you can master aperture you put into your grasp real creative control over your camera. In my opinion – aperture is where a lot of the magic happens in photography and as we’ll see below, changes in it can mean the difference between one dimensional and multi dimensional shots. What is Aperture? Put most simply – Aperture is ‘the opening in the lens.’ Cameras & Equipment When you hit the shutter release button of your camera a hole opens up that Sirui T-004X Aluminum allows your cameras image sensor to catch a glimpse of the scene you’re Tripod Review wanting to capture.
    [Show full text]
  • Modeling and Synthesis of Aperture Effects in Cameras
    Computational Aesthetics in Graphics, Visualization, and Imaging (2008) P. Brown, D. W. Cunningham, V. Interrante, and J. McCormack (Editors) Modeling and Synthesis of Aperture Effects in Cameras Douglas Lanman1,2, Ramesh Raskar1,3, and Gabriel Taubin2 1Mitsubishi Electric Research Laboratories, Cambridge, MA, (USA) 2Brown University, Division of Engineering, Providence, RI (USA) 3Massachusetts Institute of Technology, Media Lab, Cambridge, MA (USA) Abstract In this paper we describe the capture, analysis, and synthesis of optical vignetting in conventional cameras. We analyze the spatially-varying point spread function (PSF) to accurately model the vignetting for any given focus or aperture setting. In contrast to existing "flat-field" calibration procedures, we propose a simple calibration pattern consisting of a two-dimensional array of point light sources – allowing simultaneous estimation of vignetting correction tables and spatially-varying blur kernels. We demonstrate the accuracy of our model by deblurring images with focus and aperture settings not sampled during calibration. We also introduce the Bokeh Brush: a novel, post-capture method for full-resolution control of the shape of out-of-focus points. This effect is achieved by collecting a small set of images with varying basis aperture shapes. We demonstrate the effectiveness of this approach for a variety of scenes and aperture sets. Categories and Subject Descriptors (according to ACM CCS): I.3.8 [Computer Graphics]: Applications 1. Introduction 1.1. Contributions A professional photographer is faced with a seemingly great challenge: how to select the appropriate lens for a given sit- The vignetting and spatially-varying point spread function uation at a moment’s notice.
    [Show full text]
  • “No-Parallax” Point in Panorama Photography Version 1.0, February 6, 2006 Rik Littlefield ([email protected])
    Theory of the “No-Parallax” Point in Panorama Photography Version 1.0, February 6, 2006 Rik Littlefield ([email protected]) Abstract In using an ordinary camera to make panorama photographs, there is a special “no- parallax point” around which the camera must be rotated in order to keep foreground and background points lined up perfectly in overlapping frames. Arguments about the location of this point are both frequent and confusing. The purpose of this article is to eliminate the confusion. It explains in simple terms that the aperture determines the perspective of an image by selecting the light rays that form it. Therefore the center of perspective and the no-parallax point are located at the apparent position of the aperture, called the “entrance pupil”. Contrary to intuition, this point can be moved by modifying just the aperture, while leaving all refracting lens elements and the sensor in the same place. Likewise, the angle of view must be measured around the no-parallax point and thus the angle of view is affected by the aperture location, not just by the lens and sensor as is commonly believed. Physical vignetting may cause the entrance pupil to move, depending on angle, and should be avoided for perfect stitches. In general, the no- parallax point is different from the “nodal point” of optical designers, despite frequent use (misuse) of that term among panorama photographers. Introduction In using an ordinary camera to make panorama photographs, there is a special “no- parallax point”1 around which the camera must be rotated in order to keep foreground and background points lined up perfectly in overlapping frames.
    [Show full text]