Rendering Natural Camera Bokeh Effect with Deep Learning

Total Page:16

File Type:pdf, Size:1020Kb

Rendering Natural Camera Bokeh Effect with Deep Learning Rendering Natural Camera Bokeh Effect with Deep Learning Andrey Ignatov Jagruti Patel Radu Timofte [email protected] [email protected] [email protected] ETH Zurich, Switzerland Abstract Bokeh is an important artistic effect used to highlight the main object of interest on the photo by blurring all out-of-focus areas. While DSLR and system camera lenses can render this effect naturally, mobile cameras are unable to produce shallow depth-of-field photos due to a very small aperture diameter of their optics. Unlike the current solutions simulating bokeh by applying Gaussian blur to image background, in this paper we propose to learn a realistic shallow focus technique directly from the photos produced by DSLR cameras. For this, we present a large-scale bokeh dataset consisting of 5K shallow / wide depth-of-field image pairs captured using the Canon 7D DSLR with 50mm f/1.8 lenses. We use these images to train a deep learning model to reproduce a natural bokeh effect based on a single narrow-aperture image. The experimental results show that the proposed approach is able to render a plausible non-uniform bokeh even in case of complex input data with multiple objects. The dataset, pre-trained models and codes used in this paper are available on the project website: https://people.ee.ethz.ch/˜ihnatova/ pynet-bokeh.html 1. Introduction Figure 1. The original shallow depth-of-field image and the image produced with our method. Bokeh effect is a very popular photography technique used to make the subject in the shot stand out sharply unattainable for mobile cameras with compact optics and against a blurred background (Fig. 1). It is achieved by fo- tiny sensors. As a result, bokeh effect can only be sim- cusing the camera on the selected area or object and shoot- ulated computationally on smartphones and other devices ing the photo with a wide aperture lens. This produces with small mobile cameras. a shallow depth-of-field image where only objects located Synthetic bokeh effect rendering is a relatively new ma- within a narrow image plane are visible clearly, while all chine learning topic that has emerged in the past years. other parts of the image are blurred. This effect is often [15, 16] were one of the first papers exploring shallow leading to very pleasing visual results, and besides that al- depth-of-field simulation based on a single image. These lows to wash out unnecessary, distracting or unattractive works were focused on portrait photos only: they first seg- background details, which is especially useful in case of mented out people from the image with a convolutional neu- mobile photography. In order to get good bokeh, fast lenses ral network, and then uniformly blurred the remaining im- with a large aperture are needed, which makes this effect age background. The same approach was also considered 1 Figure 2. Sample wide and shallow depth-of-field image pairs from the EBB! dataset. in [21], where the authors proposed an efficient solution for In this paper, we present a different approach to the con- person segmentation. There also exists a separate group of sidered problem. Since the properties of the real bokeh de- works [8, 2, 19] that are blurring the image based on the pend on a large number of parameters such as focal length, predicted depth map, however they require the camera to distance to the subject, the acceptable circle of confusion, move in order to get the parallax effect, thus the practical the aperture size, type and construction, various optical lens applicability of these approaches is quite limited. An accu- characteristics, etc., it is almost impossible to render this rate depth map estimation can be also obtained with stereo effect precisely with the existing software tools and meth- vision [1] using synchronously captured photos from multi- ods. Therefore, in this work we propose to learn the bokeh ple cameras. technique directly from the photos produced by a high-end Since 2015, the majority of flagship and mid-range mo- DSLR camera using deep learning. The presented approach bile devices are getting special hardware for accelerating is camera independent, does not require any special hard- deep learning models [5]. The performance of such hard- ware, and can also be applied to the existing images. ware is already approaching the results of mid-range Nvidia GPUs presented in the past years [6], making the use of Our main contributions are: deep learning-based methods for bokeh effect simulation on • A novel end-to-end deep learning solution for realis- mobile devices especially relevant. Paper [17] describes the tic bokeh effect rendering. The model is trained to synthetic depth-of-field rendering approach used on Google map the standard narrow-aperture images into shallow Pixel devices that is able to achieve very impressive results depth-of-field photos captured with a DSLR camera. for the majority of photo scenes. In this work, the authors are predicting a depth map based on a single image and are • A large-scale dataset containing 5K shallow / wide using camera’s dual-pixel auto-focus system when possible depth-of-field image pairs collected in the wild with for refinement of the results. The weakest part of the paper the Canon 7D DSLR camera and 50mm f/1.8 fast lens. is the actual bokeh rendering method – a simple approxi- • A comprehensive set of experiments evaluating the mated disk blur is used for this, making the visual results quantitative and perceptual quality of the rendered quite different from the natural bokeh photos produced by bokeh photos. We demonstrate that the proposed solu- DSLR cameras. tion can process one 1024×1536 px image under 5s on all high-end mobile chipsets when running it on GPU. 2. Everything is Better with Bokeh! Dataset One of the biggest challenges in the bokeh rendering task is to get high-quality real data that can be used for training deep models. To tackle this problem, a large-scale Every- thing is Better with Bokeh! (EBB!) dataset containing more than 10 thousand images was collected in the wild during several months. By controlling the aperture size of the lens, images with shallow and wide depth-of-field were taken. In each photo pair, the first image was captured with a narrow aperture (f/16) that results in a normal sharp photo, whereas the second one was shot using the highest aperture (f/1.8) leading to a strong bokeh effect. The photos were taken during the daytime in a wide variety of places and in vari- ous illumination and weather conditions. The photos were captured in automatic mode, the default settings were used throughout the entire collection procedure. An example set of collected images is presented in Figure 2. The captured image pairs are not aligned exactly, there- fore they were first matched using SIFT keypoints and RANSAC method same as in [3]. The resulting images were then cropped to their intersection part and downscaled so that their final height is equal to 1024 pixels. Finally, we computed a coarse depth map for each wide depth-of-field image using the Megadepth model proposed in [12]. These maps can be stacked directly with the input images and used as an additional guidance for the trained model. From Figure 3. The architecture of the PyNET-based model. Concat and the resulting 10 thousand images, 200 image pairs were re- Sum ops are applied to the outputs of the adjacent layers. served for testing, while the other 4.8 thousand photo pairs can be used for training and validation. layers are then concatenated, which allows the network to 3. Proposed Method learn a more diverse set of features at each level. The out- puts obtained at lower scales are upsampled with transposed Bokeh effect simulation problem belongs to a group of convolutional layers, stacked with feature maps from the tasks dealing with both global and local image processing. upper level and then subsequently processed in the follow- High-level image analysis is needed here to detect the ar- ing convolutional layers. Leaky ReLU activation function eas on the photo where the bokeh effect should be applied, is applied after each convolutional operation, except for the whereas low-level processing is used for rendering the ac- output layers that are using tanh function to map the results tual shallow depth-of-field images and refining the results. to (-1, 1) interval. Instance normalization is used in all con- Therefore, in this work we base our solution on the PyNET volutional layers that are processing images at lower scales architecture [7] designed specifically for this kind of tasks: (levels 2-5). We are additionally using two transposed con- it is processing the image at different scales and combining volutional layers on top of the main model that upsample the learned global and local features together. the images to their target size. The model is trained sequentially, starting from the low- 3.1. PyNET CNN Architecture est layer. This allows to achieve good semantically-driven Figure 3 illustrates schematic representation of the reconstruction results at smaller scales that are working PyNET-based architecture used in this work. The model has with images of very low resolution and thus performing an inverted pyramidal shape and is processing the images mostly global image manipulations. After the bottom layer at seven different scales. The proposed architecture has a is pre-trained, the same procedure is applied to the next level number of blocks that are processing feature maps in paral- till the training is done on the original resolution.
Recommended publications
  • Depth-Aware Blending of Smoothed Images for Bokeh Effect Generation
    1 Depth-aware Blending of Smoothed Images for Bokeh Effect Generation Saikat Duttaa,∗∗ aIndian Institute of Technology Madras, Chennai, PIN-600036, India ABSTRACT Bokeh effect is used in photography to capture images where the closer objects look sharp and every- thing else stays out-of-focus. Bokeh photos are generally captured using Single Lens Reflex cameras using shallow depth-of-field. Most of the modern smartphones can take bokeh images by leveraging dual rear cameras or a good auto-focus hardware. However, for smartphones with single-rear camera without a good auto-focus hardware, we have to rely on software to generate bokeh images. This kind of system is also useful to generate bokeh effect in already captured images. In this paper, an end-to-end deep learning framework is proposed to generate high-quality bokeh effect from images. The original image and different versions of smoothed images are blended to generate Bokeh effect with the help of a monocular depth estimation network. The proposed approach is compared against a saliency detection based baseline and a number of approaches proposed in AIM 2019 Challenge on Bokeh Effect Synthesis. Extensive experiments are shown in order to understand different parts of the proposed algorithm. The network is lightweight and can process an HD image in 0.03 seconds. This approach ranked second in AIM 2019 Bokeh effect challenge-Perceptual Track. 1. Introduction tant problem in Computer Vision and has gained attention re- cently. Most of the existing approaches(Shen et al., 2016; Wad- Depth-of-field effect or Bokeh effect is often used in photog- hwa et al., 2018; Xu et al., 2018) work on human portraits by raphy to generate aesthetic pictures.
    [Show full text]
  • Depth of Field PDF Only
    Depth of Field for Digital Images Robin D. Myers Better Light, Inc. In the days before digital images, before the advent of roll film, photography was accomplished with photosensitive emulsions spread on glass plates. After processing and drying the glass negative, it was contact printed onto photosensitive paper to produce the final print. The size of the final print was the same size as the negative. During this period some of the foundational work into the science of photography was performed. One of the concepts developed was the circle of confusion. Contact prints are usually small enough that they are normally viewed at a distance of approximately 250 millimeters (about 10 inches). At this distance the human eye can resolve a detail that occupies an angle of about 1 arc minute. The eye cannot see a difference between a blurred circle and a sharp edged circle that just fills this small angle at this viewing distance. The diameter of this circle is called the circle of confusion. Converting the diameter of this circle into a size measurement, we get about 0.1 millimeters. If we assume a standard print size of 8 by 10 inches (about 200 mm by 250 mm) and divide this by the circle of confusion then an 8x10 print would represent about 2000x2500 smallest discernible points. If these points are equated to their equivalence in digital pixels, then the resolution of a 8x10 print would be about 2000x2500 pixels or about 250 pixels per inch (100 pixels per centimeter). The circle of confusion used for 4x5 film has traditionally been that of a contact print viewed at the standard 250 mm viewing distance.
    [Show full text]
  • Portraiture, Surveillance, and the Continuity Aesthetic of Blur
    Michigan Technological University Digital Commons @ Michigan Tech Michigan Tech Publications 6-22-2021 Portraiture, Surveillance, and the Continuity Aesthetic of Blur Stefka Hristova Michigan Technological University, [email protected] Follow this and additional works at: https://digitalcommons.mtu.edu/michigantech-p Part of the Arts and Humanities Commons Recommended Citation Hristova, S. (2021). Portraiture, Surveillance, and the Continuity Aesthetic of Blur. Frames Cinema Journal, 18, 59-98. http://doi.org/10.15664/fcj.v18i1.2249 Retrieved from: https://digitalcommons.mtu.edu/michigantech-p/15062 Follow this and additional works at: https://digitalcommons.mtu.edu/michigantech-p Part of the Arts and Humanities Commons Portraiture, Surveillance, and the Continuity Aesthetic of Blur Stefka Hristova DOI:10.15664/fcj.v18i1.2249 Frames Cinema Journal ISSN 2053–8812 Issue 18 (Jun 2021) http://www.framescinemajournal.com Frames Cinema Journal, Issue 18 (June 2021) Portraiture, Surveillance, and the Continuity Aesthetic of Blur Stefka Hristova Introduction With the increasing transformation of photography away from a camera-based analogue image-making process into a computerised set of procedures, the ontology of the photographic image has been challenged. Portraits in particular have become reconfigured into what Mark B. Hansen has called “digital facial images” and Mitra Azar has subsequently reworked into “algorithmic facial images.” 1 This transition has amplified the role of portraiture as a representational device, as a node in a network
    [Show full text]
  • Depth of Focus (DOF)
    Erect Image Depth of Focus (DOF) unit: mm Also known as ‘depth of field’, this is the distance (measured in the An image in which the orientations of left, right, top, bottom and direction of the optical axis) between the two planes which define the moving directions are the same as those of a workpiece on the limits of acceptable image sharpness when the microscope is focused workstage. PG on an object. As the numerical aperture (NA) increases, the depth of 46 focus becomes shallower, as shown by the expression below: λ DOF = λ = 0.55µm is often used as the reference wavelength 2·(NA)2 Field number (FN), real field of view, and monitor display magnification unit: mm Example: For an M Plan Apo 100X lens (NA = 0.7) The depth of focus of this objective is The observation range of the sample surface is determined by the diameter of the eyepiece’s field stop. The value of this diameter in 0.55µm = 0.6µm 2 x 0.72 millimeters is called the field number (FN). In contrast, the real field of view is the range on the workpiece surface when actually magnified and observed with the objective lens. Bright-field Illumination and Dark-field Illumination The real field of view can be calculated with the following formula: In brightfield illumination a full cone of light is focused by the objective on the specimen surface. This is the normal mode of viewing with an (1) The range of the workpiece that can be observed with the optical microscope. With darkfield illumination, the inner area of the microscope (diameter) light cone is blocked so that the surface is only illuminated by light FN of eyepiece Real field of view = from an oblique angle.
    [Show full text]
  • Using Depth Mapping to Realize Bokeh Effect with a Single Camera Android Device EE368 Project Report Authors (SCPD Students): Jie Gong, Ran Liu, Pradeep Vukkadala
    Using Depth Mapping to realize Bokeh effect with a single camera Android device EE368 Project Report Authors (SCPD students): Jie Gong, Ran Liu, Pradeep Vukkadala Abstract- In this paper we seek to produce a bokeh Bokeh effect is usually achieved in high end SLR effect with a single image taken from an Android device cameras using portrait lenses that are relatively large in size by post processing. Depth mapping is the core of Bokeh and have a shallow depth of field. It is extremely difficult effect production. A depth map is an estimate of depth to achieve the same effect (physically) in smart phones at each pixel in the photo which can be used to identify which have miniaturized camera lenses and sensors. portions of the image that are far away and belong to However, the latest iPhone 7 has a portrait mode which can the background and therefore apply a digital blur to the produce Bokeh effect thanks to the dual cameras background. We present algorithms to determine the configuration. To compete with iPhone 7, Google recently defocus map from a single input image. We obtain a also announced that the latest Google Pixel Phone can take sparse defocus map by calculating the ratio of gradients photos with Bokeh effect, which would be achieved by from original image and reblured image. Then, full taking 2 photos at different depths to camera and defocus map is obtained by propagating values from combining then via software. There is a gap that neither of edges to entire image by using nearest neighbor method two biggest players can achieve Bokeh effect only using a and matting Laplacian.
    [Show full text]
  • AG-AF100 28Mm Wide Lens
    Contents 1. What change when you use the different imager size camera? 1. What happens? 2. Focal Length 2. Iris (F Stop) 3. Flange Back Adjustment 2. Why Bokeh occurs? 1. F Stop 2. Circle of confusion diameter limit 3. Airy Disc 4. Bokeh by Diffraction 5. 1/3” lens Response (Example) 6. What does In/Out of Focus mean? 7. Depth of Field 8. How to use Bokeh to shoot impressive pictures. 9. Note for AF100 shooting 3. Crop Factor 1. How to use Crop Factor 2. Foal Length and Depth of Field by Imager Size 3. What is the benefit of large sensor? 4. Appendix 1. Size of Imagers 2. Color Separation Filter 3. Sensitivity Comparison 4. ASA Sensitivity 5. Depth of Field Comparison by Imager Size 6. F Stop to get the same Depth of Field 7. Back Focus and Flange Back (Flange Focal Distance) 8. Distance Error by Flange Back Error 9. View Angle Formula 10. Conceptual Schema – Relationship between Iris and Resolution 11. What’s the difference between Video Camera Lens and Still Camera Lens 12. Depth of Field Formula 1.What changes when you use the different imager size camera? 1. Focal Length changes 58mm + + It becomes 35mm Full Frame Standard Lens (CANON, NIKON, LEICA etc.) AG-AF100 28mm Wide Lens 2. Iris (F Stop) changes *distance to object:2m Depth of Field changes *Iris:F4 2m 0m F4 F2 X X <35mm Still Camera> 0.26m 0.2m 0.4m 0.26m 0.2m F4 <4/3 inch> X 0.9m X F2 0.6m 0.4m 0.26m 0.2m Depth of Field 3.
    [Show full text]
  • Adaptive Optics in Laser Processing Patrick S
    Salter and Booth Light: Science & Applications (2019) 8:110 Official journal of the CIOMP 2047-7538 https://doi.org/10.1038/s41377-019-0215-1 www.nature.com/lsa REVIEW ARTICLE Open Access Adaptive optics in laser processing Patrick S. Salter 1 and Martin J. Booth1 Abstract Adaptive optics are becoming a valuable tool for laser processing, providing enhanced functionality and flexibility for a range of systems. Using a single adaptive element, it is possible to correct for aberrations introduced when focusing inside the workpiece, tailor the focal intensity distribution for the particular fabrication task and/or provide parallelisation to reduce processing times. This is particularly promising for applications using ultrafast lasers for three- dimensional fabrication. We review recent developments in adaptive laser processing, including methods and applications, before discussing prospects for the future. Introduction enhance ultrafast DLW. An adaptive optical element Over the past two decades, direct laser writing (DLW) enables control over the fabrication laser beam and allows with ultrafast lasers has developed into a mature, diverse it to be dynamically updated during processing. Adaptive – and industrially relevant field1 5. The ultrashort nature of elements can modulate the phase, amplitude and/or the laser pulses means that energy can be delivered to the polarisation of the fabrication beam, providing many focus in a period shorter than the characteristic timescale possibilities for advanced control of the laser fabrication for thermal diffusion, leading to highly accurate material process. In this review, we briefly outline the application modification1. Thus, by focusing ultrashort pulses onto areas of AO for laser processing before considering the 1234567890():,; 1234567890():,; 1234567890():,; 1234567890():,; the surface of the workpiece, precise cuts and holes can be methods of AO, including the range of adaptive elements manufactured with a minimal heat-affected zone.
    [Show full text]
  • A Guide to Smartphone Astrophotography National Aeronautics and Space Administration
    National Aeronautics and Space Administration A Guide to Smartphone Astrophotography National Aeronautics and Space Administration A Guide to Smartphone Astrophotography A Guide to Smartphone Astrophotography Dr. Sten Odenwald NASA Space Science Education Consortium Goddard Space Flight Center Greenbelt, Maryland Cover designs and editing by Abbey Interrante Cover illustrations Front: Aurora (Elizabeth Macdonald), moon (Spencer Collins), star trails (Donald Noor), Orion nebula (Christian Harris), solar eclipse (Christopher Jones), Milky Way (Shun-Chia Yang), satellite streaks (Stanislav Kaniansky),sunspot (Michael Seeboerger-Weichselbaum),sun dogs (Billy Heather). Back: Milky Way (Gabriel Clark) Two front cover designs are provided with this book. To conserve toner, begin document printing with the second cover. This product is supported by NASA under cooperative agreement number NNH15ZDA004C. [1] Table of Contents Introduction.................................................................................................................................................... 5 How to use this book ..................................................................................................................................... 9 1.0 Light Pollution ....................................................................................................................................... 12 2.0 Cameras ................................................................................................................................................
    [Show full text]
  • Intro to Digital Photography.Pdf
    ABSTRACT Learn and master the basic features of your camera to gain better control of your photos. Individualized chapters on each of the cameras basic functions as well as cheat sheets you can download and print for use while shooting. Neuberger, Lawrence INTRO TO DGMA 3303 Digital Photography DIGITAL PHOTOGRAPHY Mastering the Basics Table of Contents Camera Controls ............................................................................................................................. 7 Camera Controls ......................................................................................................................... 7 Image Sensor .............................................................................................................................. 8 Camera Lens .............................................................................................................................. 8 Camera Modes ............................................................................................................................ 9 Built-in Flash ............................................................................................................................. 11 Viewing System ........................................................................................................................ 11 Image File Formats ....................................................................................................................... 13 File Compression ......................................................................................................................
    [Show full text]
  • The Daguerreotype Achromat 2.9/64 Art Lens by Lomography Is a Revival of This Lost Aesthetic
    OFFICIAL PRESS RELEASE The Lomographic Society Proudly Presents: Press Release Date: 6th April 2016 The DAGUERREOTYPE PRESS CONTACT Katherine Phipps [email protected] ACHROMAT 2.9/64 ART LENS Lomographic Society International 41 W 8th St The Great Return of a Lost Aesthetic from 1839 New York, NY 10011 - for Modern Day Cameras LINKS FOR EDITORS Back us on Kickstarter Now: www.lomography.com/kickstarter · The Ethereal Aesthetics of The World’s First Visit the Daguerreotype Achromat Photographic Optic Lens: reviving an optical design 2.9/64 Art Lens Microsite: www.lomography.com/achromat from 1839 by Daguerre and Chevalier · A Highly Versatile Tool for Modern Day Cameras: offering crisp sharpness, silky soft focus, freedom to control depth of field with multiple bokeh effects · Premium Quality Craftsmanship of The Art Lens Family: designed and handcrafted in a small manufactory, available in black and brass finish – reinvented for use with modern digital and analogue cameras · Proud to be Back on Kickstarter: following Lomography’s history of successful campaigns and exciting projects The Ethereal Aesthetics Of The World’s First Optic Lens Practical photography was invented in 1839, with the combination of a Chevalier Achromat Lens attached to a Daguerreotype camera. The signature character of the Chevalier lens bathed images in an alluring veil of light, due to a series of beautiful “aberrations” in its image-forming optical system, which naturally caused a glazy, soft picture at wide apertures. The Daguerreotype Achromat 2.9/64 Art Lens by Lomography is a revival of this lost aesthetic. Today, more photographs are taken every two minutes than humanity took in the entire 19th century.
    [Show full text]
  • A Curriculum Guide
    FOCUS ON PHOTOGRAPHY: A CURRICULUM GUIDE This page is an excerpt from Focus on Photography: A Curriculum Guide Written by Cynthia Way for the International Center of Photography © 2006 International Center of Photography All rights reserved. Published by the International Center of Photography, New York. Printed in the United States of America. Please credit the International Center of Photography on all reproductions. This project has been made possible with generous support from Andrew and Marina Lewin, the GE Fund, and public funds from the New York City Department of Cultural Affairs Cultural Challenge Program. FOCUS ON PHOTOGRAPHY: A CURRICULUM GUIDE PART IV Resources FOCUS ON PHOTOGRAPHY: A CURRICULUM GUIDE This section is an excerpt from Focus on Photography: A Curriculum Guide Written by Cynthia Way for the International Center of Photography © 2006 International Center of Photography All rights reserved. Published by the International Center of Photography, New York. Printed in the United States of America. Please credit the International Center of Photography on all reproductions. This project has been made possible with generous support from Andrew and Marina Lewin, the GE Fund, and public funds from the New York City Department of Cultural Affairs Cultural Challenge Program. FOCUS ON PHOTOGRAPHY: A CURRICULUM GUIDE Focus Lesson Plans Fand Actvities INDEX TO FOCUS LINKS Focus Links Lesson Plans Focus Link 1 LESSON 1: Introductory Polaroid Exercises Focus Link 2 LESSON 2: Camera as a Tool Focus Link 3 LESSON 3: Photographic Field
    [Show full text]
  • Samyang T-S 24Mm F/3.5 ED AS
    GEAR GEAR ON TEST: Samyang T-S 24mm f/3.5 ED AS UMC As the appeal of tilt-shift lenses continues to broaden, Samyang has unveiled a 24mm perspective-control optic in a range of popular fittings – and at a price that’s considerably lower than rivals from Nikon and Canon WORDS TERRY HOPE raditionally, tilt-shift lenses have been seen as specialist products, aimed at architectural T photographers wanting to correct converging verticals and product photographers seeking to maximise depth-of-field. The high price of such lenses reflects the low numbers sold, as well as the precision nature of their design and construction. But despite this, the tilt-shift lens seems to be undergoing something of a resurgence in popularity. This increasing appeal isn’t because photographers are shooting more architecture or box shots. It’s more down to the popularity of miniaturisation effects in landscapes, where such a tiny part of the frame is in focus that it appears as though you are looking at a scale model, rather than the real thing. It’s an effect that many hobbyist DSLRs, CSCs and compacts can generate digitally, and it’s straightforward to create in Photoshop too, but these digital recreations are just approximations. To get the real thing you’ll need to shoot with a tilt-shift lens. Up until now that would cost you over £1500, and when you consider the number of assignments you might shoot in a year using a tilt-shift lens, that’s not great value for money. Things are set to change, though, because there’s The optical quality of the lens is very good, and is a new kid on the block in the shape of the Samyang T-S testament to the reputation that Samyang is rapidly 24mm f/3.5 ED AS UMC, which has a street price of less gaining in the optics market.
    [Show full text]