List of Abbreviations STEM Scanning Transmission Electron Microscopy

Total Page:16

File Type:pdf, Size:1020Kb

List of Abbreviations STEM Scanning Transmission Electron Microscopy List of Abbreviations STEM Scanning Transmission Electron Microscopy PSF Point Spread Function HAADF High-Angle Annular Dark-Field Z-Contrast Atomic Number-Contrast ICM Iterative Constrained Methods RL Richardson-Lucy BD Blind Deconvolution ML Maximum Likelihood MBD Multichannel Blind Deconvolution SA Simulated Annealing SEM Scanning Electron Microscopy TEM Transmission Electron Microscopy MI Mutual Information PRF Point Response Function Registration and 3D Deconvolution of STEM Data M. MERCIMEK*, A. F. KOSCHAN*, A. Y. BORISEVICH‡, A.R. LUPINI‡, M. A. ABIDI*, & S. J. PENNYCOOK‡. *IRIS Laboratory, Electrical Engineering and Computer Science, The University of Tennessee, 1508 Middle Dr, Knoxville, TN 37996 ‡Materials Science and Technology Division, Oak Ridge National Laboratory, Oak Ridge, TN 37831 Key words. 3D deconvolution, Depth sectioning, Electron microscope, 3D Reconstruction. Summary A common characteristic of many practical modalities is that imaging process distorts the signals or scene of interest. The principal data is veiled by noise and less important signals. We split the 3D enhancement process into two separate stages; deconvolution in general is referring to the improvement of fidelity in electronic signals through reversing the convolution process, and registration aims to overcome significant misalignment of the same structures in lateral planes along focal axis. The dataset acquired with VG HB603U STEM operated at 300 kV, producing a beam with a diameter of 0.6 A˚ equipped with a Nion aberration corrector has been studied. The initial experiments described how aberration corrected Z- Contrast Imaging in STEM can be combined with the enhancement process towards better visualization with a depth resolution of a few nanometers. I. INTRODUCTION 3D characterization of both biological samples and non-biological materials provides us nano-scale resolution in three dimensions permitting the study of complex relationships between structure and existing functions. In Microscopic imaging a collection of 2D images of an object taken at different planes through optical depth sectioning is one of the bases for the reconstruction of the specimen in 3D space. Although the 3D resolution of tomography method is superior, optical depth sectioning has advantages in terms of reduction in sample irradiation, requirement of fewer images, low processing time [Nellist et al. (2006)]. As 3D images can be combined by pilling up consecutive sections, lateral views can be presented to the observer. Especially for the researchers in electron microscopy field, revealing atomic arrangements by means of visual correction of the specimen is indispensable in gaining knowledge in several ways; first principles calculations, chemical reactivity measurements, electrical properties, point defects, and optical properties. Rather than just visualizing 3D sparse point cloud, two main approximations can be adopted [Diaspro et al. (1996)]. The first approach is to extract distinctive features from the set of 2D images. The segmented images using simple binary thresholding and digital filtering can be used to produce 3D representation of the primitives. These primitives such as polygon meshes or contours are then rendered using conventional surface rendering techniques (Fig.1). The second approach is to transform the set of acquired images into another set of images such that the blurring effects are removed by investigating the image formation system and to apply rendering volumetrically afterwards. A major advantage of the latter approach is that the 3D volume can be displayed without any knowledge of the geometry of the dataset and hence without intermediate conversion to a surface representation. The entire dataset is preserved in volume rendering, any part, including internal structures and details that may be lost when reducing to geometric structures with surface rendering. In order to achieve a good representation of the structures, non-uniform behavior of the 3D transfer function of the imaging system should be corrected. Moreover during optical depth sectioning, displacements results in field of scan drifts which is mostly but not purely rigid for different focal planes. These correction methods are basically called deconvolution and registration techniques respectively and are outlined in the following sections. The purpose of this paper is to examine these two sources of degradations impairing the observed Atomic Number-contrast (Z-contrast) mode signals of STEM imaging system. Fig.1: Surface rendering towards 3D reconstruction for amorphous structure of gold particles on carbon substrate. a) Raw 2D slice image of the material from at an in-focus plane, b) Image after digital filtering and contour extraction operations, c)-d) 3D model estimate through triangulation of edge images of all slices after triangulation is applied to the distinctive edge points. II. IMAGING CHARACTERISTICS An imaging system deforms the signal of interest. It is degraded by noise, blur and the presence of other extraneous data. Separating the data stream into its useful components is essential to product improved evaluation. This linear model can precisely represent the degradation of the true image or object function, g(n1 ,n2 ,n3 ) f (n1 ,n2 ,n3 ) h(n1 ,n2 ,n3 ) n(n1 ,n2 ,n3 ) (1) where (n1, n2 , n3 )represents discrete voxel coordinates of the image frame, g(n1, n2 , n3 ) is the observed image, f (n1,n2 ,n3 ) is the true image, h(n1 , n2 , n3 ) is the image formation function or Point Spread Function (PSF), is the convolution process, n(n1 ,n2 ,n3 ) is the additive noise. For the real applications (1) can be described as a convolution integral, g(x, y, z) f (x , y , z ).h(,, ).d.d.d n(x, y, z) (2) in real world coordinates. Since in most of the physical applications, data is stored and displayed electronically, form (1) is used for computations. This model enables a simple multiplicative–additive representation in the spatial-frequency (Fourier domain), G(1,2 ,3 ) F(1,2 ,3 ) H(1,2 ,3 ) N(1,2 ,3 ) (3) Assuming that the imaging process is linear and shift-invariant a single function completely can describe the functioning of the instrument at any point in the specimen. Observed image can be considered superposition of the emitted signals (fluorescence microscopy), or transmitted signals (incoherent Z- contrast imaging in STEM) as a result of „illuminating‟ each point of the sample with PSF. The convolution essentially shifts the PSF in 3D space so that it is centered at each point in the specimen and then sums the contributions of all these shifted PSFs [McNally et all. (1999)]. Although fundamentally different, laser scanning confocal microscopy, wide-field microscopy, scanning electron microscopy, transmission electron microscopy, scanning transmission electron microscopy, aerial imaging, astronomy, are all based on the same process of image formation produced by a point source object. It can be characterized both theoretically and experimentally, and a precise understanding of the physical parameters specific to each imaging system forms the basis of most deconvolution algorithms. In electron microscopy, spherical aberration takes its name from light optics; too spherical shapes of the early lenses. Spherical aberration causes rays at higher angles to be over-focused. Chromatic aberration causes rays at different energies (different wavelengths) of electrons to be focused differently. These aberrations are difficult to describe perfectly in theoretically terms and cause a breakdown of axial symmetry and shift invariance properties. In Z-contrast mode of the STEM, the image is an incoherent one, which is simply the true image blurred by an amount independent of the specimen. Mathematically, the observed image is given in Pennycook et al. (2003) as; I(R) O(R) P 2 N(R) (4) R represents coordinates of the data, O is the object function or true image function, often represented simply as a function of Z -atomic number for each atom-, and P 2 is the effective probe intensity profile, including any broadening within the crystal. The incident electron beam is placed on one point of the sample and the electrons are scattered by the atoms are collected at the high-angle annular dark-field (HAADF) detector. The probe then moves to the next pixel and the scattered electrons are collected again at the detector. In STEM imaging as an example, a probe of atomic dimensions is scanned across a specimen. The annular detector located after specimen detects transmitted highly scattered intensity reaching, expressed as Z-contrast imaging. This kind of an image only represents the probability that the electrons strike a certain position on the detector. Electrons have wave-like properties, with wavelengths much less than that of the visible light. The de Broglie wavelength for electrons was defined as; 1.22 . (5) E where E is the energy of the electrons. Our naked-eye has ability to distinguish details up to 0.1-0.2 mm. For a 100keV electron probe the wavelength is ~0.004 nm (4 pm). Considering visible light has wavelengths between 400 nm and 700 nm, through using an electron microscope much better resolution, which defined as the smallest distance between two points that can be distinguished, can be achieved. The Rayleigh criterion describes it as; 0.61 (6) nsin where n is the refractive index of the medium surrounding the lens, and is the semi angle of the electron probe with the z axis or the acceptance angle. The small electron wavelength yields an appropriate resolution for seeing atomic structures. Major difficulties about this imaging system which we have to overcome with are; 1. In every 2D slice appears a lot of intensity information that does not represent an object appearing at this slice depth. Practically, a lot of pixels seem to have a significant value; however some of them should be removed from the scene by an algorithm. This superfluous data will make it very difficult to reconstruct edges and-or guess a threshold value that will result in useful simplified data to go on with. 2. Some of the simplest approaches to deconvolution, actually ignore the imaging formula altogether.
Recommended publications
  • Single‑Molecule‑Based Super‑Resolution Imaging
    Histochem Cell Biol (2014) 141:577–585 DOI 10.1007/s00418-014-1186-1 REVIEW The changing point‑spread function: single‑molecule‑based super‑resolution imaging Mathew H. Horrocks · Matthieu Palayret · David Klenerman · Steven F. Lee Accepted: 20 January 2014 / Published online: 11 February 2014 © Springer-Verlag Berlin Heidelberg 2014 Abstract Over the past decade, many techniques for limit, gaining over two orders of magnitude in precision imaging systems at a resolution greater than the diffraction (Szymborska et al. 2013), allowing direct observation of limit have been developed. These methods have allowed processes at spatial scales much more compatible with systems previously inaccessible to fluorescence micros- the regime that biomolecular interactions take place on. copy to be studied and biological problems to be solved in Radically, different approaches have so far been proposed, the condensed phase. This brief review explains the basic including limiting the illumination of the sample to regions principles of super-resolution imaging in both two and smaller than the diffraction limit (targeted switching and three dimensions, summarizes recent developments, and readout) or stochastically separating single fluorophores in gives examples of how these techniques have been used to time to gain resolution in space (stochastic switching and study complex biological systems. readout). The latter also described as follows: Single-mol- ecule active control microscopy (SMACM), or single-mol- Keywords Single-molecule microscopy · Super- ecule localization microscopy (SMLM), allows imaging of resolution imaging · PALM/(d)STORM imaging · single molecules which cannot only be precisely localized, Localization microscopy but also followed through time and quantified. This brief review will focus on this “pointillism-based” SR imaging and its application to biological imaging in both two and Fluorescence microscopy allows users to dynamically three dimensions.
    [Show full text]
  • Mapping the PSF Across Adaptive Optics Images
    Mapping the PSF across Adaptive Optics images Laura Schreiber Osservatorio Astronomico di Bologna Email: [email protected] Abstract • Adaptive Optics (AO) has become a key technology for all the main existing telescopes (VLT, Keck, Gemini, Subaru, LBT..) and is considered a kind of enabling technology for future giant telescopes (E-ELT, TMT, GMT). • AO increases the energy concentration of the Point Spread Function (PSF) almost reaching the resolution imposed by the diffraction limit, but the PSF itself is characterized by complex shape, no longer easily representable with an analytical model, and by sometimes significant spatial variation across the image, depending on the AO flavour and configuration. • The aim of this lesson is to describe the AO PSF characteristics and variation in order to provide (together with some AO tips) basic elements that could be useful for AO images data reduction. Erice School 2015: Science and Technology with E-ELT What’s PSF • ‘The Point Spread Function (PSF) describes the response of an imaging system to a point source’ • Circular aperture of diameter D at a wavelenght λ (no aberrations) Airy diffraction disk 2 퐼휃 = 퐼0 퐽1(푥)/푥 Where 퐽1(푥) represents the Bessel function of order 1 푥 = 휋 퐷 휆 푠푛휗 휗 is the angular radius from the aperture center First goes to 0 when 휗 ~ 1.22 휆 퐷 Erice School 2015: Science and Technology with E-ELT Imaging of a point source through a general aperture Consider a plane wave propagating in the z direction and illuminating an aperture. The element ds = dudv becomes the a sourse of a secondary spherical wave.
    [Show full text]
  • Topic 3: Operation of Simple Lens
    V N I E R U S E I T H Y Modern Optics T O H F G E R D I N B U Topic 3: Operation of Simple Lens Aim: Covers imaging of simple lens using Fresnel Diffraction, resolu- tion limits and basics of aberrations theory. Contents: 1. Phase and Pupil Functions of a lens 2. Image of Axial Point 3. Example of Round Lens 4. Diffraction limit of lens 5. Defocus 6. The Strehl Limit 7. Other Aberrations PTIC D O S G IE R L O P U P P A D E S C P I A S Properties of a Lens -1- Autumn Term R Y TM H ENT of P V N I E R U S E I T H Y Modern Optics T O H F G E R D I N B U Ray Model Simple Ray Optics gives f Image Object u v Imaging properties of 1 1 1 + = u v f The focal length is given by 1 1 1 = (n − 1) + f R1 R2 For Infinite object Phase Shift Ray Optics gives Delta Fn f Lens introduces a path length difference, or PHASE SHIFT. PTIC D O S G IE R L O P U P P A D E S C P I A S Properties of a Lens -2- Autumn Term R Y TM H ENT of P V N I E R U S E I T H Y Modern Optics T O H F G E R D I N B U Phase Function of a Lens δ1 δ2 h R2 R1 n P0 P ∆ 1 With NO lens, Phase Shift between , P0 ! P1 is 2p F = kD where k = l with lens in place, at distance h from optical, F = k0d1 + d2 +n(D − d1 − d2)1 Air Glass @ A which can be arranged to|giv{ze } | {z } F = knD − k(n − 1)(d1 + d2) where d1 and d2 depend on h, the ray height.
    [Show full text]
  • Diffraction Optics
    ECE 425 CLASS NOTES – 2000 DIFFRACTION OPTICS Physical basis • Considers the wave nature of light, unlike geometrical optics • Optical system apertures limit the extent of the wavefronts • Even a “perfect” system, from a geometrical optics viewpoint, will not form a point image of a point source • Such a system is called diffraction-limited Diffraction as a linear system • Without proof here, we state that the impulse response of diffraction is the Fourier transform (squared) of the exit pupil of the optical system (see Gaskill for derivation) 207 DR. ROBERT A. SCHOWENGERDT [email protected] 520 621-2706 (voice), 520 621-8076 (fax) ECE 425 CLASS NOTES – 2000 • impulse response called the Point Spread Function (PSF) • image of a point source • The transfer function of diffraction is the Fourier transform of the PSF • called the Optical Transfer Function (OTF) Diffraction-Limited PSF • Incoherent light, circular aperture ()2 J1 r' PSF() r' = 2-------------- r' where J1 is the Bessel function of the first kind and the normalized radius r’ is given by, 208 DR. ROBERT A. SCHOWENGERDT [email protected] 520 621-2706 (voice), 520 621-8076 (fax) ECE 425 CLASS NOTES – 2000 D = aperture diameter πD πr f = focal length r' ==------- r -------- where λf λN N = f-number λ = wavelength of light λf • The first zero occurs at r = 1.22 ------ = 1.22λN , or r'= 1.22π D Compare to the course notes on 2-D Fourier transforms 209 DR. ROBERT A. SCHOWENGERDT [email protected] 520 621-2706 (voice), 520 621-8076 (fax) ECE 425 CLASS NOTES – 2000 2-D view (contrast-enhanced) “Airy pattern” central bright region, to first- zero ring, is called the “Airy disk” 1 0.8 0.6 PSF 0.4 radial profile 0.2 0 0246810 normalized radius r' 210 DR.
    [Show full text]
  • Improved Single-Molecule Localization Precision in Astigmatism-Based 3D Superresolution
    bioRxiv preprint doi: https://doi.org/10.1101/304816; this version posted April 19, 2018. The copyright holder for this preprint (which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under aCC-BY 4.0 International license. Improved single-molecule localization precision in astigmatism-based 3D superresolution imaging using weighted likelihood estimation Christopher H. Bohrer1Ϯ, Xinxing Yang1Ϯ, Zhixin Lyu1, Shih-Chin Wang1, Jie Xiao1* 1Department of Biophysics and Biophysical Chemistry, Johns Hopkins School of Medicine, Baltimore, Maryland, 21205, USA. Ϯ These authors contributed equally to this work. *Correspondence and requests for materials should be addressed J.X. (email: [email protected]). Abstract: Astigmatism-based superresolution microscopy is widely used to determine the position of individual fluorescent emitters in three-dimensions (3D) with subdiffraction-limited resolutions. This point spread function (PSF) engineering technique utilizes a cylindrical lens to modify the shape of the PSF and break its symmetry above and below the focal plane. The resulting asymmetric PSFs at different z-positions for single emitters are fit with an elliptical 2D-Gaussian function to extract the widths along two principle x- and y-axes, which are then compared with a pre-measured calibration function to determine its z-position. While conceptually simple and easy to implement, in practice, distorted PSFs due to an imperfect optical system often compromise the localization precision; and it is laborious to optimize a multi-purpose optical system. Here we present a methodology that is independent of obtaining a perfect PSF and enhances the localization precision along the z-axis.
    [Show full text]
  • Confocal Microscopes
    Faculty of Science Debye Institute Measuring point spread functions and calibrating (STED) confocal microscopes Bachelor thesis Peter N.A. Speets Physics and Astronomy main supervisors Prof.dr. Alfons van Blaaderen Soft Condensed Matter Group Dr. Arnout Imhof Soft Condensed Matter Group daily supervisor Ernest B. van der Wee Soft Condensed Matter Group June 2016 Abstract In this thesis the point spread function (PSF) used to characterize a confocal microscope was determined for different configurations. The PSF was determined for a sample with a refractive index matched to the refractive index that the objective was designed for, which is in this case n = 1:518 and n = 1:451 [1], and it was investigated how a mismatch in refractive index affects the shape and size of the PSF. Also the PSF of stimulated emission depletion (STED) confocal microscopy was determined to investigate the advantages of STED over conventional confocal microscopy. The alignment of the STED and the excitation laser of the confocal microscope was checked by imaging the reflection of the STED and excitation laser with a gold core nanoparticle. For the calibration of the distance measurement in a sample with a mismatch in refractive index a colloidal crystal was grown to compare the interlayer distance determined by using FTIR spectroscopy and the interlayer distance determined by imaging the crystal with the confocal microscope. 2 1. INTRODUCTION 3 1 Introduction The purpose of confocal microscopy, compared to brightfield microscopy, is to increase resolution and contrast. The confocal microscope ensures that only light from the focus point in the sample is imaged.
    [Show full text]
  • Assessing Visual Quality with the Point Spread Function Using the NIDEK OPD-Scan II
    Assessing Visual Quality With the Point Spread Function Using the NIDEK OPD-Scan II Edoardo A. Ligabue, MD; Cristina Giordano, OD he measurement of high-contrast visual acuity is ABSTRACT not an adequate representation of the patient’s vi- PURPOSE: To present the use of the point spread sual quality. Visual acuity can be described by the T 1 function (PSF) metric pre- and postoperatively for the following expression : V • A • = 1/␣, where ␣ represents the assessment of visual quality in cataract and refractive minimum angle of resolution. Snellen acuity is a measure of the surgery. ability of the eye to resolve details, and the recorded value pro- vides a metric of the functional integrity of the optical system. METHODS: Case examples of cataract and refractive However, vision is a complex neural process, and the percep- surgery and the effect of accommodation are presented. All PSF measurements were obtained using the NIDEK tion of scotopic symptoms such as halos and glare cannot be OPD-Scan II, and corneal internal and total aberrations described by a single value metric. were simulated using the NIDEK OPD-Station software. Theoretically, a perfect optical system would focus all in- All eyes underwent corneal topography, wavefront aber- coming light rays into one single point that would not cause rometry, autorefraction, keratometry, and pupillometry halos or glare. In the human eye, however, light is distorted, measurements pre- and postoperatively using the OPD- Scan II. The PSF was used to assess visual quality. and this distortion can be measured using wavefront aber- rometry.2 The wave theory of light can be used to explain RESULTS: Four case examples including refractive sur- interference and the diffraction primarily responsible for gery, aspheric multifocal intraocular lens (IOL) implanta- less-than-perfect optics in the human eye that form the basis tion, toric IOL implantation, and the difference in PSF of the point spread function (PSF).3 For example, when the due to accommodation are presented.
    [Show full text]
  • An Analytical Approximation of the Hubble Space Telescope Monochromatic Point Spread Functions
    J. Astrophys. Astr. (1987) 8, 343–350 An Analytical Approximation of the Hubble Space Telescope Monochromatic Point Spread Functions Ο. Bendinelli Dipartimento di Astronomia, Università di Bologna, Via Zamboni 33, 40126 Bologna, Italy G. Parmeggiani Osservatorio Astronomico di Bologna, Via Zamboni 33, 40126 Bologna, Italy F. Zavatti Dipartimento di Astronomia, Università di Bologna, Via Zamboni 33, 40126 Bologna, Italy Received 1987 June 16; accepted 1987 October 7 Abstract. A mixture of off-centred Gaussians is used to approximate the predicted monochromatic point spread functions (PSFs) of Hubble space telescope (HST) at long wavelengths, and also, with a reduced number of parameters, the wide-band PSFs. The usefulness of the approximation in simulation and analysis of HST data is shown by means of application to image centring and convolution-deconvolution problems. Key words: HST, point spread function—image centring—image decon- volution 1, Introduction The basic information on the expected image quality of the Hubble space telescope (HST) is reported in the last Announcement of Opportunity (NASA 1984, A.O.No.OSSA-4-84, hereinafter AO). The resulting wavelength-dependent point spread functions (PSFs) are characterized (see Figs 2-E to 14-Ε in AO) by an irregular oscillatory behaviour due to superposition of the diffraction patterns of primary and secondary mirrors, a total wavefront error of 0.05 wavelength RMS (at 633 nm), and an estimated pointing instability which causes image motion with 0.007 arcsec rms. As noted by Bendinelli et al. (1985), the oscillatory behaviour at longer wavelengths does not disappear after PSF integration over pixel surface and bandwidth, though it evidently becomes smoother.
    [Show full text]
  • Psfs and Sampling of Images in Astronomy 8/25/10 4:24 PM
    PSFs and Sampling of Images in Astronomy 8/25/10 4:24 PM UP | HOME PSFs and Sampling of Images in Astronomy Already first section. (press any key to proceed) Up / HOME / HELP / toggle view Table of Contents 1 Moffat PSFs 2 The PSF as derived from models of atmospheric turbulence 3 Gaussian and multi-Gaussian PSFs 4 Airy Patterns and Diffraction Limited PSFs 5 Nyquist sampling and a band limited images 1 Moffat PSFs Moffat proposed using a softened exponential profile to model a PSF: Note that β determines the overall shape of the PSF, while α is a scale factor. The FWHM is twice the value of r at which I(r) = 0.5 I0, so derivation of the value of α corresponding to a given FWHM is straightforward. moffat <- function(r,fwhm,beta=4.765) { hwhm <- (fwhm/2.0) alpha <- hwhm*(2^(1.0/beta)-1)^(-0.5) p <- (1.0 + (r/alpha)^2.0)^(-1.0*beta) p/sum(p) } http://home.fnal.gov/~neilsen/notebook/astroPSF/astroPSF.html#undefined Page 1 of 4 PSFs and Sampling of Images in Astronomy 8/25/10 4:24 PM 2 The PSF as derived from models of atmospheric turbulence Fried (J. Opt. Soc. Am., Volume 56, p. 1372-1379) (as reported by Trujillo etal) and Woolf (1982) use atmospheric turbulence theory to show that the PSF is the Fourier transform of an exponential: where Trujillo etal find that this closely matches a Moffat PSF with β = 4.765. Typical values of β measured from astronomical images have lower values of β (Saglia etal (1993MNRAS.264..961S)), but not as low as the default value of 2.5 used by some IRAF routines.
    [Show full text]
  • Single-Molecule Localization Microscopy Hardware
    PRIMER Single-molecule localization microscopy Mickaël Lelek1,2, Melina T. Gyparaki3, Gerti Beliu4, Florian Schueder 5,6, Juliette Griffié7, Suliana Manley 7 ✉ , Ralf Jungmann 5,6 ✉ , Markus Sauer 4 ✉ , Melike Lakadamyali8,9,10 ✉ and Christophe Zimmer 1,2 ✉ Abstract | Single- molecule localization microscopy (SMLM) describes a family of powerful imaging techniques that dramatically improve spatial resolution over standard, diffraction-limited microscopy techniques and can image biological structures at the molecular scale. In SMLM, individual fluorescent molecules are computationally localized from diffraction- limited image sequences and the localizations are used to generate a super- resolution image or a time course of super- resolution images, or to define molecular trajectories. In this Primer, we introduce the basic principles of SMLM techniques before describing the main experimental considerations when performing SMLM, including fluorescent labelling, sample preparation, hardware requirements and image acquisition in fixed and live cells. We then explain how low-resolution image sequences are computationally processed to reconstruct super-resolution images and/or extract quantitative information, and highlight a selection of biological discoveries enabled by SMLM and closely related methods. We discuss some of the main limitations and potential artefacts of SMLM, as well as ways to alleviate them. Finally, we present an outlook on advanced techniques and promising new developments in the fast-evolving field of SMLM. We hope that this Primer will be a useful reference for both newcomers and practitioners of SMLM. Diffraction The spatial resolution of standard optical microscopy They have become broadly adopted in the life sciences The bending of light waves at techniques is limited to roughly half the wavelength of owing to their high spatial resolution — typically the edges of an obstacle such light.
    [Show full text]
  • Fast Point Spread Function Modeling with Deep Learning
    Prepared for submission to JCAP Fast Point Spread Function Modeling with Deep Learning Jörg Herbel,a Tomasz Kacprzak,a Adam Amara,a Alexandre Refregier,a Aurelien Lucchib aInstitute for Particle Physics and Astrophysics, Department of Physics, ETH Zurich, Wolfgang- Pauli-Strasse 27, CH-8093 Zurich, Switzerland bData Analytics Lab, Department of Computer Science, ETH Zurich, Universitätstrasse 6, CH-8092 Zurich, Switzerland E-mail: [email protected], [email protected], [email protected], [email protected], [email protected] Abstract. Modeling the Point Spread Function (PSF) of wide-field surveys is vital for many astrophysical applications and cosmological probes including weak gravitational lensing. The PSF smears the image of any recorded object and therefore needs to be taken into account when inferring properties of galaxies from astronomical images. In the case of cosmic shear, the PSF is one of the dominant sources of systematic errors and must be treated carefully to avoid biases in cosmological parameters. Recently, forward modeling approaches to calibrate shear measurements within the Monte-Carlo Control Loops (MCCL) framework have been developed. These methods typically require simulating a large amount of wide-field images, thus, the simulations need to be very fast yet have realistic properties in key features such as the PSF pattern. Hence, such forward modeling approaches require a very flexible PSF model, which is quick to evaluate and whose parameters can be estimated reliably from survey data. We present a PSF model that meets these requirements based on a fast deep-learning method to estimate its free parameters.
    [Show full text]
  • Point Spread Function of Real Wolter-I X-Ray Mirrors: Computation by Means of the Huygens-Fresnel Principle
    Point Spread Function of real Wolter-I X-ray mirrors: computation by means of the Huygens-Fresnel principle L. Raimondi1,2 and D. Spiga1 1INAF/ Osservatorio Astronomico di Brera, Via E. Bianchi 46, I-23807 Merate (LC), Italy 2Universit`adegli Studi dell’Insubria, via Valleggio, 11, I-21100, Como, Italy ABSTRACT The Point Spread Function (PSF) of an imaging X-ray telescope is mostly determined by the defects of its grazing reflection optics. In terms of Fourier components of the mirrors’ profile, figure errors comprise long spatial wavelengths, whilst short spatial wavelengths are identified as microroughness. The former contribution is in general treated by means of geometrical optics, while the contribution of the latter – strongly dependent on the X-ray energy – is usually derived from the first order scattering theory. This approach, however, requires setting a sudden transition between these two treatments, whereas in the general case the change is expectedly more gradual. In order to compute the PSF expected from a mirror profile, we need a general method that does not need to classify surface defects as roughness or profile errors. In this work we show how to compute the PSF of real X-ray mirror shells in a typical Wolter-I configuration, at any X-ray energy. The suggested method is an application of the Huygens-Fresnel principle to grazing incidence Wolter-I profiles, and is an extension of the single-reflection formalism we presented in a previous paper. A few integral equations, applied from UV to hard X-rays, return the expected PSFs from measured or simulated profiles, accounting for the surface roughness along with its PSD (Power Spectral Density).
    [Show full text]