Radiometric Corrections for Multispectral Airborne Video Imagery

Total Page:16

File Type:pdf, Size:1020Kb

Radiometric Corrections for Multispectral Airborne Video Imagery Radiometric Corrections for Multispectral Airborne Video Imagery A. Edirlsinghe, G.E. Chapman, and J.P. Louis Abstract images as well as facilitating good image mosaicking applica- Radiometric fidelity of calibrated airborne video imagery is tions. The vignetting of airborne video camera systems has been important if the imagery is to be used for quantitative analysis studied. King (1991; 1992) and Pellikka (1998)have suggested of target surfaces. A four-channel Multispectral Airborne Video some methods for correction of such vignetting. The magnitude System (MAvS) is used at Charles Sturt University (CSU)for a of vignetting-related radiometric distortions varies consider- range of environmenfal and agricultural monitoring applica- ably with operational system settings such as lens aperture. The tions. The radiometric distortions in the MAVSimagery could objective of this paper is to present the methodology and occur because of lens characteristics in the form of vignetting details of development, of a range of correction procedures to effects and optical aberrations. This paper details the control remove vignetting related radiometric distortions from the experiments conducted to detect and quantify such distortions. MAVS imagery, and to assess the accuracy of the results of these The vignetting in particular is shown to be creating a non- procedures. uniform brightness level across the MAVS imagery. The paper then develops efficient procedures to correct the vignetting in Abedom and Radlometrlc Mstoltlons In the MAVS Imagery the MAVS imagery by producing relevant correction coefticients, In practice, there can be many optical distortions arising from templates, and equations. The accumcy of these de-vignetting the camera lens geometry of a video imaging system. Generally, (correction) procedures is shown to be comparable to the ac- only a part of the light emitted by a point on the object plane, cumcy of similar corrections reported elsewhere. An investi- and which falls on the entrance aperture of the lens, emerges gation into the effect of spectrum related vignetting on the from the lens system and falls on a single point in the image MAVS imagery due to spectral filter characteristics found plane. This is true even for an ideally thin and perfectly spheri- negligible distortions that did not warrant corrections. This cal lens, where rays striking the outer portions of the lens focus particular type of vignetting is usually caused by the wave- closer to the lens than the more central rays. This is called length ships in the band-pass window of a filter at large spherical aberration. Chromatic aberration results in light of incident angles. different wavelengths being focused in different planes. Astig- matism arises where there is a different focal length in the verti- Introduction cal and the horizontal planes, and occurs when rays strike the The process of deriving quantitative rather than qualitative lens off center. Light rays striking the lens diagonally may pro- data from calibrated spaceborne or airborne sensors demands duce comet shaped focal spots known as aberration coma. In a high radiometric accuracy in addition to maintaining geomet- some cases the set of focal points corresponding to all the points ric fidelity of the imagery. For these sensor systems to achieve in the object plane do not form a perfectly flat image plane, pro- their full potential, particularly for multi-temporal studies, it is ducing an aberration called curvature of the field. important that radiometric correction procedures such as de- Due to the small MAvS field of view (FOV)(28"), and the use vignetting procedures be developed to improve the quality and of relatively smaller size apertures (fl1.7 to fl2.3) when acquir- consistency of system response. Presently, among remote sens- ing the airborne video imagery, the rays passing through the ing systems, multispectral airborne video systems are widely lenses are all close to the optical axis of the system. Therefore, used as operational remote sensing tools, due to their high spa- the use of the outer portions of the lenses is minimized and, tial and flexible spectral resolutions (Edirisinghe et al., 1999). consequently, spherical aberration, coma, astigmatism, and The four-band (blue, green, red, and near-infrared) Multispec- field curvature effects are minimal. The chromatic aberration tral Airborne Video System (MAVS) at Charles Sturt University within a band is limited to the 25-nm (narrow band) wave- (CSU) has been operationally monitoring land and water targets length difference, However, the interband chromatic aberration for a number of years (McKenzie et al., 1992;Louis et al., 1995). may be significant if the four MAVS spectral bands are not prop- A schematic diagram of the MAVS is shown in Figure 1, and the erly registered in a composite image. This can be minimized by main sensor characteristics of the MAVS (Edirisinghe et al., camera alignment, geometric correction, and digital registra- 1999;Edirisinghe, 1997) are summarized in Table 1. A 12-mm tion (Edirisinghe, 1997).Consequently, the total effects of these focal length lens is used with each of the four MAv~spectral optical aberrations on the radiometry of most of the airborne bands for normal operations. video imagery is negligible. Full radiometric correction significantly reduces image-to- Radiometric distortion is the appearance of some pixels in image brightness variation in airborne multispectral imagery, the video frame image with false brightness values causing a allowing meaningful radiometric comparison among different A. Edirisinghe is with CCMAR, CSIRO Animal Production, PMB Photogrammetric Engineering & Remote Sensing PO, Wembley, WA 6014, Australia ([email protected],au). Vol. 67, No. 8, August 2001, pp. 915-922. G.E. Chapman and J.P. Louis are with the Spatial Analysis 0099-1112/01/670&915$3.00/0 Research Group, Charles Sturt University, Wagga Wagga, NSW, O 2001 American Society for Photogrammetry 2678, Australia. and Remote Sensing PHOTOORAMMETRIC ENGINEERING & REMOTE SENSING August 200I 9s -- I light that passes through the aperture. Light fall-off is a combi- nation of two factors: the cos4 Blaw of illuminance and vignett- ing, where Bis the incident angle (Ray, 1988). The theoretical limit of the cos4 Blaw characterizes the geometrical and optical basis for the reduction of light in the periphery of the imagery according to central projection geometry (Ray, 1988). A simple derivation of the cos4 Blaw for an optical system can be found in Schreiber (1993). Light fall-off in digital imagery can analogously be mod- eled by a cos4 Blaw of illumination reduction from the center to the edge of the image, where Bis the angular deviation of a target pixel from the optical axis of the system. vignetting in an optical system occurs due to light absorp- tion by lens walls and internal shadowing of off-axis light by components of the lens (Slater, 1980). The effect of vignetting is significant for lenses with large apertures. Vignetting is the net effect of the two independent processes known as optical vignetting and mechanical vignetting (Ray, 1988). The cos4 B VGA Vista Switcher reduction of light in the periphery of the imagery is termed nat- ural vignetting (Ray, 1988). Therefore, the total effect of light fall-off can be described by the net effect of the three types of vignetting known as natural vignetting, optical vignetting, and mechanical vignetting. This description allows the term "light fall-off" to be replaced with the term "vignetting" in general. In the remainder of this paper, unless the terms natural, optical, or Figure 1. Schematic diagram of the MAVS. mechanical are used in conjunction with the word vignetting to describe specific cases of vignetting, the word vignetting in general refers to the total effect of light fall-off in the periphery of the imagery. spatially non-uniform response across the sensor FOv. In the Optical vignetting is caused by the reduction of the cross- case of the MAVS imagery, the sources of these radiometric dis- sectional area of an oblique beam traversing the lens in com- tortions are the atmosphere between the target and the sensor, parison to that of an equivalent axial beam. The effect of this is problems in the video signal digitization phase (e.g., low sig- a reduction in image illumination, due to physical length of the nal-to-noise ratio or inadequate sampling frequency rate), sun lens, the position of the aperture stop, and the diameter of the anglelview angle (reflectance) variations, cover-type bright- front and rear elements (Ray, 1988). The magnitude may be ness variation, non-uniform response due to defective sensor estimated by projecting the image of the stop and rear element elements (missing pixels) or image plane misalignment, and in the object space onto the front element. This reduction of vignetting. Generally, the full radiometric calibration of a illumination in the imagery is also known as "cat's-eye effect." remote sensing system addresses the problem of securing the The amount of vignetting depends on the f number in use; for radiometric fidelity of an imaging system. This includes radio- an aperture with high f number, the optical vignetting is mini- metric corrections such as vignetting corrections and bi-direc- mal while, for an aperture with low f number, the effect of opti- tional reflectance variation corrections, as well as establishing cal vignetting may be high. absolute radiometric calibration or building relationships Mechanical vignetting is caused by mechanical features of between radiometric response of the system and actual the lens intruding into the Fov, causing some peripheral dark- reflectancelradiance received by the sensors. This paper ening of the image due to absorption by the lens walls (Ray, addresses the problem of vignetting in the MAVS imagery and its 1988). This can be the rim of a lens hood that is too long or of correction. The radiometric calibration of the airborne video the wrong aspect ratio.
Recommended publications
  • Breaking Down the “Cosine Fourth Power Law”
    Breaking Down The “Cosine Fourth Power Law” By Ronian Siew, inopticalsolutions.com Why are the corners of the field of view in the image captured by a camera lens usually darker than the center? For one thing, camera lenses by design often introduce “vignetting” into the image, which is the deliberate clipping of rays at the corners of the field of view in order to cut away excessive lens aberrations. But, it is also known that corner areas in an image can get dark even without vignetting, due in part to the so-called “cosine fourth power law.” 1 According to this “law,” when a lens projects the image of a uniform source onto a screen, in the absence of vignetting, the illumination flux density (i.e., the optical power per unit area) across the screen from the center to the edge varies according to the fourth power of the cosine of the angle between the optic axis and the oblique ray striking the screen. Actually, optical designers know this “law” does not apply generally to all lens conditions.2 – 10 Fundamental principles of optical radiative flux transfer in lens systems allow one to tune the illumination distribution across the image by varying lens design characteristics. In this article, we take a tour into the fascinating physics governing the illumination of images in lens systems. Relative Illumination In Lens Systems In lens design, one characterizes the illumination distribution across the screen where the image resides in terms of a quantity known as the lens’ relative illumination — the ratio of the irradiance (i.e., the power per unit area) at any off-axis position of the image to the irradiance at the center of the image.
    [Show full text]
  • Sample Manuscript Showing Specifications and Style
    Information capacity: a measure of potential image quality of a digital camera Frédéric Cao 1, Frédéric Guichard, Hervé Hornung DxO Labs, 3 rue Nationale, 92100 Boulogne Billancourt, FRANCE ABSTRACT The aim of the paper is to define an objective measurement for evaluating the performance of a digital camera. The challenge is to mix different flaws involving geometry (as distortion or lateral chromatic aberrations), light (as luminance and color shading), or statistical phenomena (as noise). We introduce the concept of information capacity that accounts for all the main defects than can be observed in digital images, and that can be due either to the optics or to the sensor. The information capacity describes the potential of the camera to produce good images. In particular, digital processing can correct some flaws (like distortion). Our definition of information takes possible correction into account and the fact that processing can neither retrieve lost information nor create some. This paper extends some of our previous work where the information capacity was only defined for RAW sensors. The concept is extended for cameras with optical defects as distortion, lateral and longitudinal chromatic aberration or lens shading. Keywords: digital photography, image quality evaluation, optical aberration, information capacity, camera performance database 1. INTRODUCTION The evaluation of a digital camera is a key factor for customers, whether they are vendors or final customers. It relies on many different factors as the presence or not of some functionalities, ergonomic, price, or image quality. Each separate criterion is itself quite complex to evaluate, and depends on many different factors. The case of image quality is a good illustration of this topic.
    [Show full text]
  • Section 10 Vignetting Vignetting the Stop Determines Determines the Stop the Size of the Bundle of Rays That Propagates On-Axis an the System for Through Object
    10-1 I and Instrumentation Design Optical OPTI-502 © Copyright 2019 John E. Greivenkamp E. John 2019 © Copyright Section 10 Vignetting 10-2 I and Instrumentation Design Optical OPTI-502 Vignetting Greivenkamp E. John 2019 © Copyright On-Axis The stop determines the size of Ray Aperture the bundle of rays that propagates Bundle through the system for an on-axis object. As the object height increases, z one of the other apertures in the system (such as a lens clear aperture) may limit part or all of Stop the bundle of rays. This is known as vignetting. Vignetted Off-Axis Ray Rays Bundle z Stop Aperture 10-3 I and Instrumentation Design Optical OPTI-502 Ray Bundle – On-Axis Greivenkamp E. John 2018 © Copyright The ray bundle for an on-axis object is a rotationally-symmetric spindle made up of sections of right circular cones. Each cone section is defined by the pupil and the object or image point in that optical space. The individual cone sections match up at the surfaces and elements. Stop Pupil y y y z y = 0 At any z, the cross section of the bundle is circular, and the radius of the bundle is the marginal ray value. The ray bundle is centered on the optical axis. 10-4 I and Instrumentation Design Optical OPTI-502 Ray Bundle – Off Axis Greivenkamp E. John 2019 © Copyright For an off-axis object point, the ray bundle skews, and is comprised of sections of skew circular cones which are still defined by the pupil and object or image point in that optical space.
    [Show full text]
  • Multi-Agent Recognition System Based on Object Based Image Analysis Using Worldview-2
    intervals, between −3 and +3, for a total of 19 different EVs To mimic airborne imagery collection from a terrestrial captured (see Table 1). WB variable images were captured location, we attempted to approximate an equivalent path utilizing the camera presets listed in Table 1 at the EVs of 0, radiance to match a range of altitudes (750 to 1500 m) above −1 and +1, for a total of 27 images. Variable light metering ground level (AGL), that we commonly use as platform alti- tests were conducted using the “center weighted” light meter tudes for our simulated damage assessment research. Optical setting, at the EVs listed in Table 1, for a total of seven images. depth increases horizontally along with the zenith angle, from The variable ISO tests were carried out at the EVs of −1, 0, and a factor of one at zenith angle 0°, to around 40° at zenith angle +1, using the ISO values listed in Table 1, for a total of 18 im- 90° (Allen, 1973), and vertically, with a zenith angle of 0°, the ages. The variable aperture tests were conducted using 19 dif- magnitude of an object at the top of the atmosphere decreases ferent apertures ranging from f/2 to f/16, utilizing a different by 0.28 at sea level, 0.24 at 500 m, and 0.21 at 1,000 m above camera system from all the other tests, with ISO = 50, f = 105 sea level, to 0 at around 100 km (Green, 1992). With 25 mm, four different degrees of onboard vignetting control, and percent of atmospheric scattering due to atmosphere located EVs of −⅓, ⅔, 0, and +⅔, for a total of 228 images.
    [Show full text]
  • Holographic Optics for Thin and Lightweight Virtual Reality
    Holographic Optics for Thin and Lightweight Virtual Reality ANDREW MAIMONE, Facebook Reality Labs JUNREN WANG, Facebook Reality Labs Fig. 1. Left: Photo of full color holographic display in benchtop form factor. Center: Prototype VR display in sunglasses-like form factor with display thickness of 8.9 mm. Driving electronics and light sources are external. Right: Photo of content displayed on prototype in center image. Car scenes by komba/Shutterstock. We present a class of display designs combining holographic optics, direc- small text near the limit of human visual acuity. This use case also tional backlighting, laser illumination, and polarization-based optical folding brings VR out of the home and in to work and public spaces where to achieve thin, lightweight, and high performance near-eye displays for socially acceptable sunglasses and eyeglasses form factors prevail. virtual reality. Several design alternatives are proposed, compared, and ex- VR has made good progress in the past few years, and entirely perimentally validated as prototypes. Using only thin, flat films as optical self-contained head-worn systems are now commercially available. components, we demonstrate VR displays with thicknesses of less than 9 However, current headsets still have box-like form factors and pro- mm, fields of view of over 90◦ horizontally, and form factors approach- ing sunglasses. In a benchtop form factor, we also demonstrate a full color vide only a fraction of the resolution of the human eye. Emerging display using wavelength-multiplexed holographic lenses that uses laser optical design techniques, such as polarization-based optical folding, illumination to provide a large gamut and highly saturated color.
    [Show full text]
  • Aperture Efficiency and Wide Field-Of-View Optical Systems Mark R
    Aperture Efficiency and Wide Field-of-View Optical Systems Mark R. Ackermann, Sandia National Laboratories Rex R. Kiziah, USAF Academy John T. McGraw and Peter C. Zimmer, J.T. McGraw & Associates Abstract Wide field-of-view optical systems are currently finding significant use for applications ranging from exoplanet search to space situational awareness. Systems ranging from small camera lenses to the 8.4-meter Large Synoptic Survey Telescope are designed to image large areas of the sky with increased search rate and scientific utility. An interesting issue with wide-field systems is the known compromises in aperture efficiency. They either use only a fraction of the available aperture or have optical elements with diameters larger than the optical aperture of the system. In either case, the complete aperture of the largest optical component is not fully utilized for any given field point within an image. System costs are driven by optical diameter (not aperture), focal length, optical complexity, and field-of-view. It is important to understand the optical design trade space and how cost, performance, and physical characteristics are influenced by various observing requirements. This paper examines the aperture efficiency of refracting and reflecting systems with one, two and three mirrors. Copyright © 2018 Advanced Maui Optical and Space Surveillance Technologies Conference (AMOS) – www.amostech.com Introduction Optical systems exhibit an apparent falloff of image intensity from the center to edge of the image field as seen in Figure 1. This phenomenon, known as vignetting, results when the entrance pupil is viewed from an off-axis point in either object or image space.
    [Show full text]
  • Optics II: Practical Photographic Lenses CS 178, Spring 2013
    Optics II: practical photographic lenses CS 178, Spring 2013 Begun 4/11/13, finished 4/16/13. Marc Levoy Computer Science Department Stanford University Outline ✦ why study lenses? ✦ thin lenses • graphical constructions, algebraic formulae ✦ thick lenses • center of perspective, 3D perspective transformations ✦ depth of field ✦ aberrations & distortion ✦ vignetting, glare, and other lens artifacts ✦ diffraction and lens quality ✦ special lenses • 2 telephoto, zoom © Marc Levoy Lens aberrations ✦ chromatic aberrations ✦ Seidel aberrations, a.k.a. 3rd order aberrations • arise because we use spherical lenses instead of hyperbolic • can be modeled by adding 3rd order terms to Taylor series ⎛ φ 3 φ 5 φ 7 ⎞ sin φ ≈ φ − + − + ... ⎝⎜ 3! 5! 7! ⎠⎟ • oblique aberrations • field curvature • distortion 3 © Marc Levoy Dispersion (wikipedia) ✦ index of refraction varies with wavelength • higher dispersion means more variation • amount of variation depends on material • index is typically higher for blue than red • so blue light bends more 4 © Marc Levoy red and blue have Chromatic aberration the same focal length (wikipedia) ✦ dispersion causes focal length to vary with wavelength • for convex lens, blue focal length is shorter ✦ correct using achromatic doublet • strong positive lens + weak negative lens = weak positive compound lens • by adjusting dispersions, can correct at two wavelengths 5 © Marc Levoy The chromatic aberrations (Smith) ✦ longitudinal (axial) chromatic aberration • different colors focus at different depths • appears everywhere
    [Show full text]
  • Projection Crystallography and Micrography with Digital Cameras Andrew Davidhazy
    Rochester Institute of Technology RIT Scholar Works Articles 2004 Projection crystallography and micrography with digital cameras Andrew Davidhazy Follow this and additional works at: http://scholarworks.rit.edu/article Recommended Citation Davidhazy, Andrew, "Projection crystallography and micrography with digital cameras" (2004). Accessed from http://scholarworks.rit.edu/article/468 This Technical Report is brought to you for free and open access by RIT Scholar Works. It has been accepted for inclusion in Articles by an authorized administrator of RIT Scholar Works. For more information, please contact [email protected]. Projection Crystallography and Micrography with Digital Cameras Andrew Davidhazy Imaging and Photographic Technology Department School of Photographic Arts and Sciences Rochester Institute of Technology During a lunchroom conversation many years ago a good friend of mine, by the name of Nile Root, and I observed that a microscope is very similar to a photographic enlarger and we proceeded to introduce this concept to our classes. Eventually a student of mine by the name of Alan Pomplun wrote and article about the technique of using an enlarger as the basis for photographing crystals that exhibited birefringence. The reasoning was that in some applications a low magnification of crystalline subjects typically examined in crystallograohy was acceptable and that the light beam in an enlarger could be easily polarized above a sample placed in the negative stage. A second polarizer placed below the lens would allow the colorful patterns that some crystalline structures achieve to be come visible. His article, that was published in Industrial Photography magazine in 1988, can be retrieved here. With the advent of digital cameras the process of using an enlarger as an improvised low-power transmission microscope merits to be revisited as there are many benefits to be derived from using the digital process for the capture of images under such a "microscope" but there are also potential pitfalls to be aware of.
    [Show full text]
  • Part 4: Filed and Aperture, Stop, Pupil, Vignetting Herbert Gross
    Optical Engineering Part 4: Filed and aperture, stop, pupil, vignetting Herbert Gross Summer term 2020 www.iap.uni-jena.de 2 Contents . Definition of field and aperture . Numerical aperture and F-number . Stop and pupil . Vignetting Optical system stop . Pupil stop defines: 1. chief ray angle w 2. aperture cone angle u . The chief ray gives the center line of the oblique ray cone of an off-axis object point . The coma rays limit the off-axis ray cone . The marginal rays limit the axial ray cone stop pupil coma ray marginal ray y field angle image w u object aperture angle y' chief ray 4 Definition of Aperture and Field . Imaging on axis: circular / rotational symmetry Only spherical aberration and chromatical aberrations . Finite field size, object point off-axis: - chief ray as reference - skew ray bundels: y y' coma and distortion y p y'p - Vignetting, cone of ray bundle not circular O' field of marginal/rim symmetric view ray RAP' chief ray - to distinguish: u w' u' w tangential and sagittal chief ray plane O entrance object exit image pupil plane pupil plane 5 Numerical Aperture and F-number image . Classical measure for the opening: plane numerical aperture (half cone angle) object in infinity NA' nsinu' DEnP/2 u' . In particular for camera lenses with object at infinity: F-number f f F# DEnP . Numerical aperture and F-number are to system properties, they are related to a conjugate object/image location 1 . Paraxial relation F # 2n'tanu' . Special case for small angles or sine-condition corrected systems 1 F # 2NA' 6 Generalized F-Number .
    [Show full text]
  • Single-Image Vignetting Correction Using Radial Gradient Symmetry
    Single-Image Vignetting Correction Using Radial Gradient Symmetry Yuanjie Zheng1 Jingyi Yu1 Sing Bing Kang2 Stephen Lin3 Chandra Kambhamettu1 1 University of Delaware, Newark, DE, USA {zheng,yu,chandra}@eecis.udel.edu 2 Microsoft Research, Redmond, WA, USA [email protected] 3 Microsoft Research Asia, Beijing, P.R. China [email protected] Abstract lenge is to differentiate the global intensity variation of vi- gnetting from those caused by local texture and lighting. In this paper, we present a novel single-image vignetting Zheng et al.[25] treat intensity variation caused by texture method based on the symmetric distribution of the radial as “noise”; as such, they require some form of robust outlier gradient (RG). The radial gradient is the image gradient rejection in fitting the vignetting function. They also require along the radial direction with respect to the image center. segmentation and must explicitly account for local shading. We show that the RG distribution for natural images with- All these are susceptible to errors. out vignetting is generally symmetric. However, this dis- We are also interested in vignetting correction using a tribution is skewed by vignetting. We develop two variants single image. Our proposed approach is fundamentally dif- of this technique, both of which remove vignetting by min- ferent from Zheng et al.’s—we rely on the property of sym- imizing asymmetry of the RG distribution. Compared with metry of the radial gradient distribution. (By radial gradient, prior approaches to single-image vignetting correction, our we mean the gradient along the radial direction with respect method does not require segmentation and the results are to the image center.) We show that the radial gradient dis- generally better.
    [Show full text]
  • (DSLR) Cameras
    Optimizing Radiometric Fidelity to Enhance Aerial Image Change Detection Utilizing Digital Single Lens Reflex (DSLR) Cameras Andrew D. Kerr and Douglas A. Stow Abstract Our objectives are to analyze the radiometric characteris- benefits of replicating solar ephemeris (Coulter et al., 2012, tics and best practices for maximizing radiometric fidel- Ahrends et al., 2008), specific shutter and aperture settings ity of digital single lens reflex (DSLR) cameras for aerial (Ahrends et al., 2008, Lebourgeois et al., 2008), using RAW image-based change detection. Control settings, exposure files (Deanet al., 2000, Coulter et al., 2012, Ahrends et al., values, white balance, light metering, ISO, and lens aper- 2008, Lebourgeois et al., 2008), vignetting abatement (Dean ture are evaluated for several bi-temporal imagery datasets. et al., 2000), and maintaining intra-frame white balance (WB) These variables are compared for their effects on dynamic consistency (Richardson et al., 2009, Levin et al., 2005). range, intra-frame brightness variation, acuity, temporal Through this study we seek to identify and determine how consistency, and detectability of simulated cracks. Test- to compensate and account for, the photometric aspects of im- ing was conducted from a terrestrial, rather than airborne age capture and postprocessing with DSLR cameras, to achieve platform, due to the large number of images collected, and high radiometric fidelity within and between digital multi- to minimize inter-image misregistration. Results point to temporal images. The overall goal is to minimize the effects of exposure biases in the range of −0.7 or −0.3EV (i.e., slightly these factors on the radiometric consistency of multi-temporal less than the auto-exposure selected levels) being preferable images, inter-frame brightness, and capture of images with for change detection and noise minimization, by achiev- high acuity and dynamic range.
    [Show full text]
  • Vignetting Effect in Fourier Ptychographic Microscopy
    Vignetting effect in Fourier ptychographic microscopy 1,2 3,* 4 1,2 1 1,* AN PAN, CHAO ZUO, YUEGE XIE, YAN ZHANG, MING LEI, BAOLI YAO 1 State Key Laboratory of Transient Optics and Photonics, Xi’an Institute of Optics and Precision Mechanics, Chinese Academy of Sciences, Xi’an 710119, China 2 University of Chinese Academy of Sciences, Beijing 100049, China 3 Smart Computational Imaging (SCI) Laboratory, Nanjing University of Science and Technology, Nanjing 210094, China 4 Institute for Computational Engineering and Sciences (ICES), University of Texas at Austin, Austin, TX 78712, USA *Corresponding author: [email protected] (C. Zuo), [email protected] (B. Yao) Fourier ptychographic microscopy (FPM) is a object-image relationship [13, 14]. In real experimental conditions, computational imaging technique that overcomes the however, the resolution may decrease from the center to the edge physical space-bandwidth product (SBP) limit of a of FOV [15]. Several unexpected “semi-bright and semi-dark” conventional microscope by applying angular diversity imperfect images with strong vignetting effect can be easily illuminations. In the usual model of FPM, the microscopic observed when the sample is illuminated by the LED within the system is approximated by being taken as space- “transition zone” between bright field (BF) and dark field (DF). invariant with transfer function determined by a These imperfect images, apparently, are not coincident with the complex pupil function of the objective. However, in real LSI model and could deteriorate the reconstruction quality of FPM experimental conditions, several unexpected “semi- severely. Especially the edge of FOV, there will be obvious wrinkles.
    [Show full text]