Modeling and Synthesis of Aperture Effects in Cameras

Total Page:16

File Type:pdf, Size:1020Kb

Modeling and Synthesis of Aperture Effects in Cameras Computational Aesthetics in Graphics, Visualization, and Imaging (2008) P. Brown, D. W. Cunningham, V. Interrante, and J. McCormack (Editors) Modeling and Synthesis of Aperture Effects in Cameras Douglas Lanman1,2, Ramesh Raskar1,3, and Gabriel Taubin2 1Mitsubishi Electric Research Laboratories, Cambridge, MA, (USA) 2Brown University, Division of Engineering, Providence, RI (USA) 3Massachusetts Institute of Technology, Media Lab, Cambridge, MA (USA) Abstract In this paper we describe the capture, analysis, and synthesis of optical vignetting in conventional cameras. We analyze the spatially-varying point spread function (PSF) to accurately model the vignetting for any given focus or aperture setting. In contrast to existing "flat-field" calibration procedures, we propose a simple calibration pattern consisting of a two-dimensional array of point light sources – allowing simultaneous estimation of vignetting correction tables and spatially-varying blur kernels. We demonstrate the accuracy of our model by deblurring images with focus and aperture settings not sampled during calibration. We also introduce the Bokeh Brush: a novel, post-capture method for full-resolution control of the shape of out-of-focus points. This effect is achieved by collecting a small set of images with varying basis aperture shapes. We demonstrate the effectiveness of this approach for a variety of scenes and aperture sets. Categories and Subject Descriptors (according to ACM CCS): I.3.8 [Computer Graphics]: Applications 1. Introduction 1.1. Contributions A professional photographer is faced with a seemingly great challenge: how to select the appropriate lens for a given sit- The vignetting and spatially-varying point spread function uation at a moment’s notice. While there are a variety of capture, analysis, and synthesis methods introduced in this heuristics, such as the well-known “sunny f/16” rule, a pho- paper integrate enhancements to a number of prior results in tographer’s skill in this task must be honed by experience. It a novel way. The primary contributions include: is one of the goals of computational photography to reduce some of these concerns for both professional and amateur i. By exploiting the simple observation that the out-of-focus photographers. While previous works have examined meth- image of a point light directly gives the point spread func- ods for refocusing, deblurring, or augmenting conventional tion, we show a practical low-cost method to simulta- images, few have examined the topic of bokeh. In general, a neously estimate the vignetting and the spatially-varying good bokeh is characterized by a subtle blur for out-of-focus point spread function. (Note that, while straightforward, points – creating a pleasing separation between foreground this method can prove challenging in practice due to the and background objects in portrait or macro photography. In long exposure times required with point sources.) this paper we develop a new method to allow post-capture control of lens bokeh for still life scenes. ii. We introduce the Bokeh Brush: a novel, post-capture method for full-resolution control of the shape of out-of- To inform our discussion of image bokeh, we present a focus points. This effect is achieved by collecting a small unified approach to vignetting calibration in conventional set of images with varying basis aperture shapes. We cameras. Drawing upon recent work in computer vision and demonstrate that optimal basis aperture selection is es- graphics, we propose a simple, yet accurate, vignetting and sentially a compression problem – one solution of which spatially-varying point spread function model. This model is to apply PCA or NMF to training aperture images. and calibration procedure should find broad applicability as (Note that we assume a static scene so that multiple ex- more researchers begin exploring the topics of vignetting, posures can be obtained with varying aperture shapes.) highlight manipulation, and aesthetics. °c The Eurographics Association 2008. D. Lanman, R. Raskar, and G. Taubin / Modeling and Synthesis of Aperture Effects in Cameras 1.2. Related Work The topic of vignetting correction can be subsumed within the larger field of radiometric calibration. As described in Litvinov and Schechner [LS05], cameras exhibit three pri- mary types of radiometric non-idealities: (1) spatial non- uniformity due to vignetting, (2) nonlinear radiometric re- sponse of the sensor, and (3) temporal variations due to auto- matic gain control (AGC). Unfortunately, typical consumer- grade cameras do not allow users to precisely control intrin- sic camera parameters and settings (e.g., zoom, focal length, and aperture). As a result, laboratory flat-field calibration us- ing a uniform white light area source [Yu04] proves prob- lematic – motivating recent efforts to develop simpler radio- metric calibration procedures. Several authors have focused on single-image radiometric calibration, as well as single- image vignetting correction [ZLK06]. In most of these Figure 1: Illustration of optical vignetting. From left to works the motivating application is creating image mosaics, right: (a) reference image at f/5.6, (b) reference image at whether using a sequence of still images [GC05,d’A07] or a f/1.4, and (inset) illustration of entrance pupil shape as a video sequence [LS05]. function of incidence angle and aperture setting [vW07]. Recently, several applications in computer vision and graphics have required high-accuracy estimates of Optical vignetting occurs in multi-element optical de- spatially-varying point spread functions. Veeraraghavan et ∗ signs. As shown in Figure 1, for a given aperture the al. [VRA 07] and Levin et al. [LFDF07] considered coded clear area will decrease for off-axis viewing angles and aperture imaging. In those works, a spatially-modulated can be modeled using the variable cone model described mask (i.e., an aperture pattern) was placed at the iris plane in [AAB96]. Optical vignetting can be reduced by stopping of a conventional camera. In the former work, a broadband down the lens (i.e., reducing the aperture size), since this will mask enabled post-processing digital refocusing (at full reduce exit pupil variation for large viewing angles. sensor resolution) for layered Lambertian scenes. In the later work, the authors proposed a similar mask for simulta- Natural vignetting causes radial attenuation that, unlike neously recovering scene depth and high-resolution images. the previous types, does not arise from occlusion of light. In both cases, the authors proposed specific PSF calibration Instead, this source of vignetting is due to the physical prop- erties of light and the geometric construction of typical cam- patterns, including: general scenes under natural image 4 statistics and a planar pattern of random curves, respectively. eras. Typically modeled using the approximate cos (θ) law, We also recognize the closely-related work on confocal where θ is the angle of light leaving the rear of the lens, natu- stereo and variable-aperture photography developed by ral vignetting combines the effects due to the inverse square Hasinoff and Kutulakos [HK06, HK07]. Note that we will fall-off of light, Lambert’s law, and the foreshortening of the discuss their models in more detail in Section 2.1. exit pupil for large incidence angles [KW00, vW07]. Pixel vignetting arises in digital cameras. Similar to me- 2. Modeling Vignetting chanical and optical vignetting, pixel vignetting causes a ra- dial falloff of light recorded by a digital sensor (e.g., CMOS) Images produced by optical photography tend to exhibit a due to the finite depth of the photon well, causing light to be radial reduction in brightness that increases towards the im- blocked from the detector regions for large incidence angles. age periphery. This reduction arises from a combination of factors, including: (1) limitations of the optical design of the camera, (2) the physical properties of light, and (3) particular 2.1. Geometric Model: Spatially-varying PSF characteristics of the imaging sensor. In this work we sepa- Recall that a thin lens can be characterized by rate these effects using the taxonomy presented by Goldman 1 1 1 and Chen [GC05] and Ray [Ray02]. = + , f fD D Mechanical vignetting results in radial brightness atten- where f is the focal length, f is the separation between uation due to physical obstructions in front of the lens body. D the image and lens planes, and D is the distance to the Typical obstructions include lens hoods, filters, and sec- object plane (see Figure 2). As described by Bae and Du- ondary lenses. In contrast to other types of vignetting, me- rand [BD07], the diameter c of the PSF is given by chanical vignetting can completely block light from reach- ing certain image regions, preventing those areas from being |S − D| f 2 c = · , recovered by any correction algorithm. S N(D − f ) °c The Eurographics Association 2008. D. Lanman, R. Raskar, and G. Taubin / Modeling and Synthesis of Aperture Effects in Cameras where S is the distance to a given out-of-focus point and N is the f-number. This model predicts that the PSF will scale as a function of the object distance S and the f-number N. As a result, a calibration procedure would need to sample both these parameters to fully characterize the point spread func- tion. However, as noted by Hasinoff and Kutulakos [HK07], the effective blur diameterc ˜ is given by the linear relation |S − D| c˜ = A, S where A is the aperture diameter. Under this approximation, we find that the spatially varying PSF could potentially be estimated from a single image. In conclusion, we find that the spatially-varying PSF B(s,t;x,y) will scale linearly with Figure 3: Example of piecewise-linear point spread function the effective blur diameterc ˜ such that interpolation. Gray kernels correspond to images of blurred 1 x y point sources, whereas the red kernel is linearly interpolated Bc˜(s,t;x,y)= Bc˜ s,t; , , c˜2 ³ c˜ c˜´ from its three nearest neighbors.
Recommended publications
  • Depth-Aware Blending of Smoothed Images for Bokeh Effect Generation
    1 Depth-aware Blending of Smoothed Images for Bokeh Effect Generation Saikat Duttaa,∗∗ aIndian Institute of Technology Madras, Chennai, PIN-600036, India ABSTRACT Bokeh effect is used in photography to capture images where the closer objects look sharp and every- thing else stays out-of-focus. Bokeh photos are generally captured using Single Lens Reflex cameras using shallow depth-of-field. Most of the modern smartphones can take bokeh images by leveraging dual rear cameras or a good auto-focus hardware. However, for smartphones with single-rear camera without a good auto-focus hardware, we have to rely on software to generate bokeh images. This kind of system is also useful to generate bokeh effect in already captured images. In this paper, an end-to-end deep learning framework is proposed to generate high-quality bokeh effect from images. The original image and different versions of smoothed images are blended to generate Bokeh effect with the help of a monocular depth estimation network. The proposed approach is compared against a saliency detection based baseline and a number of approaches proposed in AIM 2019 Challenge on Bokeh Effect Synthesis. Extensive experiments are shown in order to understand different parts of the proposed algorithm. The network is lightweight and can process an HD image in 0.03 seconds. This approach ranked second in AIM 2019 Bokeh effect challenge-Perceptual Track. 1. Introduction tant problem in Computer Vision and has gained attention re- cently. Most of the existing approaches(Shen et al., 2016; Wad- Depth-of-field effect or Bokeh effect is often used in photog- hwa et al., 2018; Xu et al., 2018) work on human portraits by raphy to generate aesthetic pictures.
    [Show full text]
  • Portraiture, Surveillance, and the Continuity Aesthetic of Blur
    Michigan Technological University Digital Commons @ Michigan Tech Michigan Tech Publications 6-22-2021 Portraiture, Surveillance, and the Continuity Aesthetic of Blur Stefka Hristova Michigan Technological University, [email protected] Follow this and additional works at: https://digitalcommons.mtu.edu/michigantech-p Part of the Arts and Humanities Commons Recommended Citation Hristova, S. (2021). Portraiture, Surveillance, and the Continuity Aesthetic of Blur. Frames Cinema Journal, 18, 59-98. http://doi.org/10.15664/fcj.v18i1.2249 Retrieved from: https://digitalcommons.mtu.edu/michigantech-p/15062 Follow this and additional works at: https://digitalcommons.mtu.edu/michigantech-p Part of the Arts and Humanities Commons Portraiture, Surveillance, and the Continuity Aesthetic of Blur Stefka Hristova DOI:10.15664/fcj.v18i1.2249 Frames Cinema Journal ISSN 2053–8812 Issue 18 (Jun 2021) http://www.framescinemajournal.com Frames Cinema Journal, Issue 18 (June 2021) Portraiture, Surveillance, and the Continuity Aesthetic of Blur Stefka Hristova Introduction With the increasing transformation of photography away from a camera-based analogue image-making process into a computerised set of procedures, the ontology of the photographic image has been challenged. Portraits in particular have become reconfigured into what Mark B. Hansen has called “digital facial images” and Mitra Azar has subsequently reworked into “algorithmic facial images.” 1 This transition has amplified the role of portraiture as a representational device, as a node in a network
    [Show full text]
  • Using Depth Mapping to Realize Bokeh Effect with a Single Camera Android Device EE368 Project Report Authors (SCPD Students): Jie Gong, Ran Liu, Pradeep Vukkadala
    Using Depth Mapping to realize Bokeh effect with a single camera Android device EE368 Project Report Authors (SCPD students): Jie Gong, Ran Liu, Pradeep Vukkadala Abstract- In this paper we seek to produce a bokeh Bokeh effect is usually achieved in high end SLR effect with a single image taken from an Android device cameras using portrait lenses that are relatively large in size by post processing. Depth mapping is the core of Bokeh and have a shallow depth of field. It is extremely difficult effect production. A depth map is an estimate of depth to achieve the same effect (physically) in smart phones at each pixel in the photo which can be used to identify which have miniaturized camera lenses and sensors. portions of the image that are far away and belong to However, the latest iPhone 7 has a portrait mode which can the background and therefore apply a digital blur to the produce Bokeh effect thanks to the dual cameras background. We present algorithms to determine the configuration. To compete with iPhone 7, Google recently defocus map from a single input image. We obtain a also announced that the latest Google Pixel Phone can take sparse defocus map by calculating the ratio of gradients photos with Bokeh effect, which would be achieved by from original image and reblured image. Then, full taking 2 photos at different depths to camera and defocus map is obtained by propagating values from combining then via software. There is a gap that neither of edges to entire image by using nearest neighbor method two biggest players can achieve Bokeh effect only using a and matting Laplacian.
    [Show full text]
  • AG-AF100 28Mm Wide Lens
    Contents 1. What change when you use the different imager size camera? 1. What happens? 2. Focal Length 2. Iris (F Stop) 3. Flange Back Adjustment 2. Why Bokeh occurs? 1. F Stop 2. Circle of confusion diameter limit 3. Airy Disc 4. Bokeh by Diffraction 5. 1/3” lens Response (Example) 6. What does In/Out of Focus mean? 7. Depth of Field 8. How to use Bokeh to shoot impressive pictures. 9. Note for AF100 shooting 3. Crop Factor 1. How to use Crop Factor 2. Foal Length and Depth of Field by Imager Size 3. What is the benefit of large sensor? 4. Appendix 1. Size of Imagers 2. Color Separation Filter 3. Sensitivity Comparison 4. ASA Sensitivity 5. Depth of Field Comparison by Imager Size 6. F Stop to get the same Depth of Field 7. Back Focus and Flange Back (Flange Focal Distance) 8. Distance Error by Flange Back Error 9. View Angle Formula 10. Conceptual Schema – Relationship between Iris and Resolution 11. What’s the difference between Video Camera Lens and Still Camera Lens 12. Depth of Field Formula 1.What changes when you use the different imager size camera? 1. Focal Length changes 58mm + + It becomes 35mm Full Frame Standard Lens (CANON, NIKON, LEICA etc.) AG-AF100 28mm Wide Lens 2. Iris (F Stop) changes *distance to object:2m Depth of Field changes *Iris:F4 2m 0m F4 F2 X X <35mm Still Camera> 0.26m 0.2m 0.4m 0.26m 0.2m F4 <4/3 inch> X 0.9m X F2 0.6m 0.4m 0.26m 0.2m Depth of Field 3.
    [Show full text]
  • Intro to Digital Photography.Pdf
    ABSTRACT Learn and master the basic features of your camera to gain better control of your photos. Individualized chapters on each of the cameras basic functions as well as cheat sheets you can download and print for use while shooting. Neuberger, Lawrence INTRO TO DGMA 3303 Digital Photography DIGITAL PHOTOGRAPHY Mastering the Basics Table of Contents Camera Controls ............................................................................................................................. 7 Camera Controls ......................................................................................................................... 7 Image Sensor .............................................................................................................................. 8 Camera Lens .............................................................................................................................. 8 Camera Modes ............................................................................................................................ 9 Built-in Flash ............................................................................................................................. 11 Viewing System ........................................................................................................................ 11 Image File Formats ....................................................................................................................... 13 File Compression ......................................................................................................................
    [Show full text]
  • The Daguerreotype Achromat 2.9/64 Art Lens by Lomography Is a Revival of This Lost Aesthetic
    OFFICIAL PRESS RELEASE The Lomographic Society Proudly Presents: Press Release Date: 6th April 2016 The DAGUERREOTYPE PRESS CONTACT Katherine Phipps [email protected] ACHROMAT 2.9/64 ART LENS Lomographic Society International 41 W 8th St The Great Return of a Lost Aesthetic from 1839 New York, NY 10011 - for Modern Day Cameras LINKS FOR EDITORS Back us on Kickstarter Now: www.lomography.com/kickstarter · The Ethereal Aesthetics of The World’s First Visit the Daguerreotype Achromat Photographic Optic Lens: reviving an optical design 2.9/64 Art Lens Microsite: www.lomography.com/achromat from 1839 by Daguerre and Chevalier · A Highly Versatile Tool for Modern Day Cameras: offering crisp sharpness, silky soft focus, freedom to control depth of field with multiple bokeh effects · Premium Quality Craftsmanship of The Art Lens Family: designed and handcrafted in a small manufactory, available in black and brass finish – reinvented for use with modern digital and analogue cameras · Proud to be Back on Kickstarter: following Lomography’s history of successful campaigns and exciting projects The Ethereal Aesthetics Of The World’s First Optic Lens Practical photography was invented in 1839, with the combination of a Chevalier Achromat Lens attached to a Daguerreotype camera. The signature character of the Chevalier lens bathed images in an alluring veil of light, due to a series of beautiful “aberrations” in its image-forming optical system, which naturally caused a glazy, soft picture at wide apertures. The Daguerreotype Achromat 2.9/64 Art Lens by Lomography is a revival of this lost aesthetic. Today, more photographs are taken every two minutes than humanity took in the entire 19th century.
    [Show full text]
  • Three Techniques for Rendering Generalized Depth of Field Effects
    Three Techniques for Rendering Generalized Depth of Field Effects Todd J. Kosloff∗ Computer Science Division, University of California, Berkeley, CA 94720 Brian A. Barskyy Computer Science Division and School of Optometry, University of California, Berkeley, CA 94720 Abstract Post-process methods are fast, sometimes to the point Depth of field refers to the swath that is imaged in sufficient of real-time [13, 17, 9], but generally do not share the focus through an optics system, such as a camera lens. same image quality as distributed ray tracing. Control over depth of field is an important artistic tool that can be used to emphasize the subject of a photograph. In a A full literature review of depth of field methods real camera, the control over depth of field is limited by the is beyond the scope of this paper, but the interested laws of physics and by physical constraints. Depth of field reader should consult the following surveys: [1, 2, 5]. has been rendered in computer graphics, but usually with the same limited control as found in real camera lenses. In Kosara [8] introduced the notion of semantic depth this paper, we generalize depth of field in computer graphics of field, a somewhat similar notion to generalized depth by allowing the user to specify the distribution of blur of field. Semantic depth of field is non-photorealistic throughout a scene in a more flexible manner. Generalized depth of field provides a novel tool to emphasize an area of depth of field used for visualization purposes. Semantic interest within a 3D scene, to select objects from a crowd, depth of field operates at a per-object granularity, and to render a busy, complex picture more understandable allowing each object to have a different amount of blur.
    [Show full text]
  • Photography Techniques Intermediate Skills
    Photography Techniques Intermediate Skills PDF generated using the open source mwlib toolkit. See http://code.pediapress.com/ for more information. PDF generated at: Wed, 21 Aug 2013 16:20:56 UTC Contents Articles Bokeh 1 Macro photography 5 Fill flash 12 Light painting 12 Panning (camera) 15 Star trail 17 Time-lapse photography 19 Panoramic photography 27 Cross processing 33 Tilted plane focus 34 Harris shutter 37 References Article Sources and Contributors 38 Image Sources, Licenses and Contributors 39 Article Licenses License 41 Bokeh 1 Bokeh In photography, bokeh (Originally /ˈboʊkɛ/,[1] /ˈboʊkeɪ/ BOH-kay — [] also sometimes heard as /ˈboʊkə/ BOH-kə, Japanese: [boke]) is the blur,[2][3] or the aesthetic quality of the blur,[][4][5] in out-of-focus areas of an image. Bokeh has been defined as "the way the lens renders out-of-focus points of light".[6] However, differences in lens aberrations and aperture shape cause some lens designs to blur the image in a way that is pleasing to the eye, while others produce blurring that is unpleasant or distracting—"good" and "bad" bokeh, respectively.[2] Bokeh occurs for parts of the scene that lie outside the Coarse bokeh on a photo shot with an 85 mm lens and 70 mm entrance pupil diameter, which depth of field. Photographers sometimes deliberately use a shallow corresponds to f/1.2 focus technique to create images with prominent out-of-focus regions. Bokeh is often most visible around small background highlights, such as specular reflections and light sources, which is why it is often associated with such areas.[2] However, bokeh is not limited to highlights; blur occurs in all out-of-focus regions of the image.
    [Show full text]
  • Backlighting
    Challenge #1 - Capture Light Bokeh Have you seen those beautiful little circles of colour and light in the background of photos? It’s called Bokeh and it comes from the Japanese word “boke” meaning blur. It’s so pretty and not only does it make a gorgeous backdrop for portraits, but it can be the subject in it’s own right too! It’s also one of the most fun aspects of learning photography, being able to capture your own lovely bokeh! And that’s what our first challenge is all about, Capturing Light Bokeh! And remember - just as I said in this video, the purpose of these challenges is NOT to take perfect photos the first time you try something new… it’s about looking for light, trying new techniques, and exploring your creativity to help you build your skills and motivate and inspire you. So lets have fun & I can’t wait to see your photos! Copyright 2017 - www.clicklovegrow.com How to Achieve Light Bokeh Light Bokeh is created when light reflects off, or through, a background; and is then captured with a wide open aperture of your lens. As light reflects differently off flat surfaces, you’re not likely to see this effect when using a plain wall, for example. Instead, you’ll see it when your background has a little texture, such as light reflecting off leaves in foliage of a garden, or wet grass… or when light is broken up when streaming through trees. Looking for Light One of the key take-aways for this challenge is to start paying attention to light & how you can capture it in photos… both to add interest and feeling to your portraits, but as a creative subject in it’s own right.
    [Show full text]
  • Introduction to Aperture in Digital Photography Helpful
    9/7/2015 Introduction to Aperture - Digital Photography School Tips & Tutorials Cameras & Equipment Post Production Books Presets Courses Forums Search Introduction to Aperture in Join 1,417,294 Subscribers! Digital Photography A Post By: Darren Rowse 142K 23.1K 661 164 1200 96 0 35 Shares Share Share Share Pin it Share Other Comments Photography Tips & Tutorials Tips for Getting Started with Urban Landscape Photography 6 Tips for Creating More Captivating Landscape Photographs How to Solve 5 Composition Tempest by Sheldon Evans on 500px Conundrums Faced by Over the last couple of weeks I’ve been writing a series of posts on elements Landscape Photographers that digital photographers need to learn about in order to get out of Auto mode and learn how to manually set the exposure of their shots. I’ve largely focussed Landscape Photography - upon three elements of the ‘exposure triangle‘ – ISO, Shutter Speed and Shooting the Same Aperture. I’ve previously written about the first two and today would like to Location Through the turn our attention to Aperture. Seasons Before I start with the explanations let me say this. If you can master aperture you put into your grasp real creative control over your camera. In my opinion – aperture is where a lot of the magic happens in photography and as we’ll see below, changes in it can mean the difference between one dimensional and multi dimensional shots. What is Aperture? Put most simply – Aperture is ‘the opening in the lens.’ Cameras & Equipment When you hit the shutter release button of your camera a hole opens up that Sirui T-004X Aluminum allows your cameras image sensor to catch a glimpse of the scene you’re Tripod Review wanting to capture.
    [Show full text]
  • “No-Parallax” Point in Panorama Photography Version 1.0, February 6, 2006 Rik Littlefield ([email protected])
    Theory of the “No-Parallax” Point in Panorama Photography Version 1.0, February 6, 2006 Rik Littlefield ([email protected]) Abstract In using an ordinary camera to make panorama photographs, there is a special “no- parallax point” around which the camera must be rotated in order to keep foreground and background points lined up perfectly in overlapping frames. Arguments about the location of this point are both frequent and confusing. The purpose of this article is to eliminate the confusion. It explains in simple terms that the aperture determines the perspective of an image by selecting the light rays that form it. Therefore the center of perspective and the no-parallax point are located at the apparent position of the aperture, called the “entrance pupil”. Contrary to intuition, this point can be moved by modifying just the aperture, while leaving all refracting lens elements and the sensor in the same place. Likewise, the angle of view must be measured around the no-parallax point and thus the angle of view is affected by the aperture location, not just by the lens and sensor as is commonly believed. Physical vignetting may cause the entrance pupil to move, depending on angle, and should be avoided for perfect stitches. In general, the no- parallax point is different from the “nodal point” of optical designers, despite frequent use (misuse) of that term among panorama photographers. Introduction In using an ordinary camera to make panorama photographs, there is a special “no- parallax point”1 around which the camera must be rotated in order to keep foreground and background points lined up perfectly in overlapping frames.
    [Show full text]
  • Art Lenses (Beginner Level) Vintage and Art Lenses
    Creative Portraits Day 2 - Vintage and Art Lenses (Beginner Level) Vintage and Art Lenses There are many ways to create interest and intrigue in a portrait that extends beyond simply getting the sharpest and cleanest looking image you can. In fact we can produce creative portraits with only the lens itself and in this instance I'm talking about either vintage lenses or art lenses. The images above are taken using 'art' lenses that create a very distinctive look. Vintage Lenses A vintage lens is really just an older lens that doesn't have all of the modern optics that we all familiar with now. For example they will have features like a stronger vignette, unique looking bokeh or soft focusing towards the edges etc. Not only are there so many different lenses available to us today but we've also been making lens for well over a hundred years so there are so many older lenses we've never used but they're all unique. I've had a lot of fun playing with these vintage lenses and the great part is that they are generally incredibly cheap to buy. You should be able to pick up some vintage lenses for as cheap as £30/$50 on certain auction sites. The only downside is that they are all going to There are literally hundreds of vintage lenses available to us today. be old enough that they won't talk to your camera, so auto-focus isn't available, you can't operate you camera in automatic mode in fact nearly all the modern features are unavailable but that's ok because the shots you'll be taking will look fantastically unique and creative.
    [Show full text]