Depth of Focus By
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Still Photography
Still Photography Soumik Mitra, Published by - Jharkhand Rai University Subject: STILL PHOTOGRAPHY Credits: 4 SYLLABUS Introduction to Photography Beginning of Photography; People who shaped up Photography. Camera; Lenses & Accessories - I What a Camera; Types of Camera; TLR; APS & Digital Cameras; Single-Lens Reflex Cameras. Camera; Lenses & Accessories - II Photographic Lenses; Using Different Lenses; Filters. Exposure & Light Understanding Exposure; Exposure in Practical Use. Photogram Introduction; Making Photogram. Darkroom Practice Introduction to Basic Printing; Photographic Papers; Chemicals for Printing. Suggested Readings: 1. Still Photography: the Problematic Model, Lew Thomas, Peter D'Agostino, NFS Press. 2. Images of Information: Still Photography in the Social Sciences, Jon Wagner, 3. Photographic Tools for Teachers: Still Photography, Roy A. Frye. Introduction to Photography STILL PHOTOGRAPHY Course Descriptions The department of Photography at the IFT offers a provocative and experimental curriculum in the setting of a large, diversified university. As one of the pioneers programs of graduate and undergraduate study in photography in the India , we aim at providing the best to our students to help them relate practical studies in art & craft in professional context. The Photography program combines the teaching of craft, history, and contemporary ideas with the critical examination of conventional forms of art making. The curriculum at IFT is designed to give students the technical training and aesthetic awareness to develop a strong individual expression as an artist. The faculty represents a broad range of interests and aesthetics, with course offerings often reflecting their individual passions and concerns. In this fundamental course, students will identify basic photographic tools and their intended purposes, including the proper use of various camera systems, light meters and film selection. -
Depth-Aware Blending of Smoothed Images for Bokeh Effect Generation
1 Depth-aware Blending of Smoothed Images for Bokeh Effect Generation Saikat Duttaa,∗∗ aIndian Institute of Technology Madras, Chennai, PIN-600036, India ABSTRACT Bokeh effect is used in photography to capture images where the closer objects look sharp and every- thing else stays out-of-focus. Bokeh photos are generally captured using Single Lens Reflex cameras using shallow depth-of-field. Most of the modern smartphones can take bokeh images by leveraging dual rear cameras or a good auto-focus hardware. However, for smartphones with single-rear camera without a good auto-focus hardware, we have to rely on software to generate bokeh images. This kind of system is also useful to generate bokeh effect in already captured images. In this paper, an end-to-end deep learning framework is proposed to generate high-quality bokeh effect from images. The original image and different versions of smoothed images are blended to generate Bokeh effect with the help of a monocular depth estimation network. The proposed approach is compared against a saliency detection based baseline and a number of approaches proposed in AIM 2019 Challenge on Bokeh Effect Synthesis. Extensive experiments are shown in order to understand different parts of the proposed algorithm. The network is lightweight and can process an HD image in 0.03 seconds. This approach ranked second in AIM 2019 Bokeh effect challenge-Perceptual Track. 1. Introduction tant problem in Computer Vision and has gained attention re- cently. Most of the existing approaches(Shen et al., 2016; Wad- Depth-of-field effect or Bokeh effect is often used in photog- hwa et al., 2018; Xu et al., 2018) work on human portraits by raphy to generate aesthetic pictures. -
Depth of Focus (DOF)
Erect Image Depth of Focus (DOF) unit: mm Also known as ‘depth of field’, this is the distance (measured in the An image in which the orientations of left, right, top, bottom and direction of the optical axis) between the two planes which define the moving directions are the same as those of a workpiece on the limits of acceptable image sharpness when the microscope is focused workstage. PG on an object. As the numerical aperture (NA) increases, the depth of 46 focus becomes shallower, as shown by the expression below: λ DOF = λ = 0.55µm is often used as the reference wavelength 2·(NA)2 Field number (FN), real field of view, and monitor display magnification unit: mm Example: For an M Plan Apo 100X lens (NA = 0.7) The depth of focus of this objective is The observation range of the sample surface is determined by the diameter of the eyepiece’s field stop. The value of this diameter in 0.55µm = 0.6µm 2 x 0.72 millimeters is called the field number (FN). In contrast, the real field of view is the range on the workpiece surface when actually magnified and observed with the objective lens. Bright-field Illumination and Dark-field Illumination The real field of view can be calculated with the following formula: In brightfield illumination a full cone of light is focused by the objective on the specimen surface. This is the normal mode of viewing with an (1) The range of the workpiece that can be observed with the optical microscope. With darkfield illumination, the inner area of the microscope (diameter) light cone is blocked so that the surface is only illuminated by light FN of eyepiece Real field of view = from an oblique angle. -
Adaptive Optics in Laser Processing Patrick S
Salter and Booth Light: Science & Applications (2019) 8:110 Official journal of the CIOMP 2047-7538 https://doi.org/10.1038/s41377-019-0215-1 www.nature.com/lsa REVIEW ARTICLE Open Access Adaptive optics in laser processing Patrick S. Salter 1 and Martin J. Booth1 Abstract Adaptive optics are becoming a valuable tool for laser processing, providing enhanced functionality and flexibility for a range of systems. Using a single adaptive element, it is possible to correct for aberrations introduced when focusing inside the workpiece, tailor the focal intensity distribution for the particular fabrication task and/or provide parallelisation to reduce processing times. This is particularly promising for applications using ultrafast lasers for three- dimensional fabrication. We review recent developments in adaptive laser processing, including methods and applications, before discussing prospects for the future. Introduction enhance ultrafast DLW. An adaptive optical element Over the past two decades, direct laser writing (DLW) enables control over the fabrication laser beam and allows with ultrafast lasers has developed into a mature, diverse it to be dynamically updated during processing. Adaptive – and industrially relevant field1 5. The ultrashort nature of elements can modulate the phase, amplitude and/or the laser pulses means that energy can be delivered to the polarisation of the fabrication beam, providing many focus in a period shorter than the characteristic timescale possibilities for advanced control of the laser fabrication for thermal diffusion, leading to highly accurate material process. In this review, we briefly outline the application modification1. Thus, by focusing ultrashort pulses onto areas of AO for laser processing before considering the 1234567890():,; 1234567890():,; 1234567890():,; 1234567890():,; the surface of the workpiece, precise cuts and holes can be methods of AO, including the range of adaptive elements manufactured with a minimal heat-affected zone. -
A Curriculum Guide
FOCUS ON PHOTOGRAPHY: A CURRICULUM GUIDE This page is an excerpt from Focus on Photography: A Curriculum Guide Written by Cynthia Way for the International Center of Photography © 2006 International Center of Photography All rights reserved. Published by the International Center of Photography, New York. Printed in the United States of America. Please credit the International Center of Photography on all reproductions. This project has been made possible with generous support from Andrew and Marina Lewin, the GE Fund, and public funds from the New York City Department of Cultural Affairs Cultural Challenge Program. FOCUS ON PHOTOGRAPHY: A CURRICULUM GUIDE PART IV Resources FOCUS ON PHOTOGRAPHY: A CURRICULUM GUIDE This section is an excerpt from Focus on Photography: A Curriculum Guide Written by Cynthia Way for the International Center of Photography © 2006 International Center of Photography All rights reserved. Published by the International Center of Photography, New York. Printed in the United States of America. Please credit the International Center of Photography on all reproductions. This project has been made possible with generous support from Andrew and Marina Lewin, the GE Fund, and public funds from the New York City Department of Cultural Affairs Cultural Challenge Program. FOCUS ON PHOTOGRAPHY: A CURRICULUM GUIDE Focus Lesson Plans Fand Actvities INDEX TO FOCUS LINKS Focus Links Lesson Plans Focus Link 1 LESSON 1: Introductory Polaroid Exercises Focus Link 2 LESSON 2: Camera as a Tool Focus Link 3 LESSON 3: Photographic Field -
Depth of Field and Depth of Focus Explained
1 Reference: Queries. Vol. 31 /1 Proceedings RMS March 1996 pp64-66 Depth of Field and depth of Focus Explained What is the difference between 'depth of field' and 'depth of focus'? - I have seen both terms used in textbooks This is a good question, which prompts an answer that encompasses more than simply terminology. In the past there has been some confusion about these terms, and they have been used rather indiscriminately. Recently, usage has become more consistent, and when we published the RMS Dictionary of light Microscopy in 1989, we felt able to make an unambiguous distinction in our somewhat ponderous definitions: • Depth of field is the axial depth of the space on both sides of the object plane within which the object can be moved without detectable loss of sharpness in the image, and within which features of the object appear acceptably sharp in the image while the position of the image plane is maintained. • Depth of focus is the axial depth of the space on both sides of the image plane within which the image appears acceptably sharp while the positions of the object plane and of the objective are maintained. The essential distinction between the terms is clear: depth of field refers to object space and depth of focus to image space. A possibly useful mnemonic is that the field of view is that part of the object that is being examined, and the focus is the point at which parallel rays converge after passing through a lens. These definitions contain two rather imprecise terms: detectable loss of sharpness, and acceptably sharp; clearly it all depends on what can be detected and what is considered acceptable. -
A Theoretical and Practical Introduction to Optics a Theoretical and Practical Introduction to Optics
White Paper A Theoretical and Practical Introduction to Optics A Theoretical and Practical Introduction to Optics Be honest: do you really know how to calculate the focal length of a lens? If so, you are an exception to the rule and can stop reading here !! For the rest of you, here is a second chance. Back to square one "Piece of broken glass starts forest fire"– a common headline during the summer. But how could this have happened? Due to the enormous distance between the Earth and the Sun, the Sun only appears as a tiny point emitting parallel rays of light (figure 1a) Should these parallel rays pass through a lens (or a piece of glass, which has similar characteristics) the rays would meet behind the lens at what is called the focal point. But what happens if our point of light is so near to the lens that we can not assume to have parallel rays of light? They cross each other behind the focal point (figure 1b). If we take a look at the image of our point of light at the focal points position we will see a unclear blurred spot. And so the question arises- "what is focusing?". Focusing is to increase the distance between the focal plane and the lens until the focal plane and the junction of the rays overlap each other (figure 1c). Thus, for single points of light the situation is quite simple. But what happens to the image of screws, PCBs or plates of steel? From points of light to images A point of light does not necessarily originate directly from the sun, candles or lamps, it can also result from a reflection. -
Mission to the Solar Gravity Lens Focus: Natural Highground for Imaging Earth-Like Exoplanets L
Planetary Science Vision 2050 Workshop 2017 (LPI Contrib. No. 1989) 8203.pdf MISSION TO THE SOLAR GRAVITY LENS FOCUS: NATURAL HIGHGROUND FOR IMAGING EARTH-LIKE EXOPLANETS L. Alkalai1, N. Arora1, M. Shao1, S. Turyshev1, L. Friedman8 ,P. C. Brandt3, R. McNutt3, G. Hallinan2, R. Mewaldt2, J. Bock2, M. Brown2, J. McGuire1, A. Biswas1, P. Liewer1, N. Murphy1, M. Desai4, D. McComas5, M. Opher6, E. Stone2, G. Zank7, 1Jet Propulsion Laboratory, Pasadena, CA 91109, USA, 2California Institute of Technology, Pasadena, CA 91125, USA, 3The Johns Hopkins University Applied Physics Laboratory, Laurel, MD 20723, USA,4Southwest Research Institute, San Antonio, TX 78238, USA, 5Princeton Plasma Physics Laboratory, Princeton, NJ 08543, USA, 6Boston University, Boston, MA 02215, USA, 7University of Alabama in Huntsville, Huntsville, AL 35899, USA, 8Emritus, The Planetary Society. Figure 1: A SGL Probe Mission is a first step in the goal to search and study potential habitable exoplanets. This figure was developed as a product of two Keck Institute for Space Studies (KISS) workshops on the topic of the “Science and Enabling Technologies for the Exploration of the Interstellar Medium” led by E. Stone, L. Alkalai and L. Friedman. Introduction: Recent data from Voyager 1, Kepler and New Horizons spacecraft have resulted in breath-taking discoveries that have excited the public and invigorated the space science community. Voyager 1, the first spacecraft to arrive at the Heliopause, discovered that Fig. 2. Imaging of an exo-Earth with solar gravitational Lens. The exo-Earth occupies (1km×1km) area at the image plane. Using a 1m the interstellar medium is far more complicated and telescope as a 1 pixel detector provides a (1000×1000) pixel image! turbulent than expected; the Kepler telescope According to Einstein’s general relativity, gravity discovered that exoplanets are not only ubiquitous but induces refractive properties of space-time causing a also diverse in our galaxy and that Earth-like exoplanets massive object to act as a lens by bending light. -
Camera Settings Guide
Camera Settings Guide • " " and " " are trademarks or registered trademarks of Sony Corporation. • All other company and product names mentioned herein are used for identification purposes only and may be the trademarks or registered trademarks of their respective owners. TM and ® symbols are not included in this booklet. • Screen displays and effects used to illustrate some functions are simulated. Conventional autofocus has until now dealt with space alone. Sony goes one step further — a big step, with an innovative image sensor that picks up both space and time to capture moving subjects with new clarity. Sony spells the beginning of a new autofocus era. 4D FOCUS allows you to take crisper photos than ever. Plain old autofocus is a thing of the past. The future of photography is in motion. What is 4D FOCUS? Space: 3D Time: 4D 4D FOCUS Area Depth Time Wide Fast Steadfast The wide AF area, covering nearly the Fast Hybrid AF, combining phase- An advanced AF algorithm accurately entire frame, allows focusing on a detection AF and contrast-detection AF, predicts subject’s next move. Precise AF subject positioned even off the center instantly detects distance to the subject tracking allows focus to be maintained of the frame. to focus accurately. even on fast-moving subjects. Meeting your focusing demands Basic AF performance of Wide Fast Steadfast Focusing over wide area Instant focusing! Once it's focused, it never lets go The 6000 employs a focal plane phase- Advanced Fast Hybrid AF combines phase- With Focus Mode set to AF-C, the camera detection AF sensor with 179 AF points spread detection AF and contrast-detection AF to achieve displays one or more small green frames to cover nearly the entire frame. -
Digital Light Field Photography
DIGITAL LIGHT FIELD PHOTOGRAPHY a dissertation submitted to the department of computer science and the committee on graduate studies of stanford university in partial fulfillment of the requirements for the degree of doctor of philosophy Ren Ng July © Copyright by Ren Ng All Rights Reserved ii IcertifythatIhavereadthisdissertationandthat,inmyopinion,itisfully adequateinscopeandqualityasadissertationforthedegreeofDoctorof Philosophy. Patrick Hanrahan Principal Adviser IcertifythatIhavereadthisdissertationandthat,inmyopinion,itisfully adequateinscopeandqualityasadissertationforthedegreeofDoctorof Philosophy. Marc Levoy IcertifythatIhavereadthisdissertationandthat,inmyopinion,itisfully adequateinscopeandqualityasadissertationforthedegreeofDoctorof Philosophy. Mark Horowitz Approved for the University Committee on Graduate Studies. iii iv Acknowledgments I feel tremendously lucky to have had the opportunity to work with Pat Hanrahan, Marc Levoy and Mark Horowitz on the ideas in this dissertation, and I would like to thank them for their support. Pat instilled in me a love for simulating the flow of light, agreed to take me on as a graduate student, and encouraged me to immerse myself in something I had a passion for.Icouldnothaveaskedforafinermentor.MarcLevoyistheonewhooriginallydrewme to computer graphics, has worked side by side with me at the optical bench, and is vigorously carrying these ideas to new frontiers in light field microscopy. Mark Horowitz inspired me to assemble my camera by sharing his love for dismantling old things and building new ones. I have never met a professor more generous with his time and experience. I am grateful to Brian Wandell and Dwight Nishimura for serving on my orals commit- tee. Dwight has been an unfailing source of encouragement during my time at Stanford. I would like to acknowledge the fine work of the other individuals who have contributed to this camera research. Mathieu Brédif worked closely with me in developing the simulation system, and he implemented the original lens correction software. -
Longitudinal Chromatic Aberration
Longitudinal Chromatic Aberration Red focus Blue focus LCA Transverse Chromatic Aberration Decentered pupil Transverse Chromatic Aberration Chromatic Difference of Refraction Atchison and Smith, JOSA A, 2005 Why is LCA not really a problem? Chromatic Aberration Halos (LCA) Fringes (TCA) www.starizona.com digitaldailydose.wordpress.com Red-Green Duochrome test • If the letters on the red side stand out more, add minus power; if the letters on the green side stand out more, add plus power. • Neutrality is reached when the letters on both backgrounds appear equally distinct. Colligon-Bradley P. J Ophthalmic Nurs Technol. 1992 11(5):220-2. Transverse Chromatic Aberration Lab #7 April 15th Look at a red and blue target through a 1 mm pinhole. Move the pinhole from one edge of the pupil to the other. What happens to the red and blue images? Chromatic Difference of Magnification Chief ray Aperture stop Off axis source Abbe Number • Also known as – Refractive efficiency – nu-value –V-value –constringence Refractive efficiencies for common materials water 55.6 alcohol 60.6 ophthalmic crown glass 58.6 polycarbonate 30.0 dense flint glass 36.6 Highlite glass 31.0 BK7 64.9 Example • Given the following indices of refraction for BK7 glass (nD = 1.519; nF = 1.522; nC = 1.514) what is the refractive efficiency? • What is the chromatic aberration of a 20D thin lens made of BK7 glass? Problem • Design a 10.00 D achromatic doublet using ophthalmic crown glass and dense flint glass Carl Friedrich Gauss 1777-1855 Heinrich Seidel 1842-1906 Approximations to -
Computer Vision: Projection
Image formation and cameras Photography CSE P 576 Larry Zitnick ([email protected]) Many slides courtesy of Steve Seitz Image formation Pinhole camera Object Film Object Barrier Film Add a barrier to block off most of the rays • This reduces blurring Let’s design a camera • The opening known as the aperture • Idea 1: put a piece of film in front of an object • How does this transform the image? • Do we get a reasonable image? 1 Camera Obscura Shrinking the aperture The first camera • Known to Aristotle • How does the aperture size affect the image? Why not make the aperture as small as possible? • Less light gets through • Diffraction effects... Shrinking the aperture Adding a lens Object Lens Film ―circle of confusion‖ A lens focuses light onto the film • There is a specific distance at which objects are ―in focus‖ – other points project to a ―circle of confusion‖ in the image • Changing the shape of the lens changes this distance 2 Lenses Thin lenses Object Lens Film Lens Aperture Film Focal Focal point point Optical center (Center of projection) Focal length A lens focuses parallel rays onto a single focal point Not quite right… Thin lens equation: • focal point at a distance f beyond the plane of the lens – f is a function of the shape and index of refraction of the lens • Aperture of diameter D restricts the range of rays • Any object point satisfying this equation is in focus – aperture may be on either side of the lens • What is the shape of the focus region? • Lenses are typically spherical (easier to produce) • How can we change the focus region? • Thin lens applet: http://www.phy.ntnu.edu.tw/java/Lens/lens_e.html (by Fu-Kwun Hwang ) Thin lens assumption Depth of field Aperture Film The thin lens assumption assumes the lens has no thickness, but this isn’t true… Object Lens Film f / 5.6 Focal point f / 32 Changing the aperture size affects depth of field By adding more elements to the lens, the distance at which a scene is in focus can be made roughly planar.