Digital Signal Processing
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Convolution! (CDT-14) Luciano Da Fontoura Costa
Convolution! (CDT-14) Luciano da Fontoura Costa To cite this version: Luciano da Fontoura Costa. Convolution! (CDT-14). 2019. hal-02334910 HAL Id: hal-02334910 https://hal.archives-ouvertes.fr/hal-02334910 Preprint submitted on 27 Oct 2019 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Convolution! (CDT-14) Luciano da Fontoura Costa [email protected] S~aoCarlos Institute of Physics { DFCM/USP October 22, 2019 Abstract The convolution between two functions, yielding a third function, is a particularly important concept in several areas including physics, engineering, statistics, and mathematics, to name but a few. Yet, it is not often so easy to be conceptually understood, as a consequence of its seemingly intricate definition. In this text, we develop a conceptual framework aimed at hopefully providing a more complete and integrated conceptual understanding of this important operation. In particular, we adopt an alternative graphical interpretation in the time domain, allowing the shift implied in the convolution to proceed over free variable instead of the additive inverse of this variable. In addition, we discuss two possible conceptual interpretations of the convolution of two functions as: (i) the `blending' of these functions, and (ii) as a quantification of `matching' between those functions. -
High Frequency Integrated MOS Filters C
N94-71118 2nd NASA SERC Symposium on VLSI Design 1990 5.4.1 High Frequency Integrated MOS Filters C. Peterson International Microelectronic Products San Jose, CA 95134 Abstract- Several techniques exist for implementing integrated MOS filters. These techniques fit into the general categories of sampled and tuned continuous- time niters. Advantages and limitations of each approach are discussed. This paper focuses primarily on the high frequency capabilities of MOS integrated filters. 1 Introduction The use of MOS in the design of analog integrated circuits continues to grow. Competitive pressures drive system designers towards system-level integration. System-level integration typically requires interface to the analog world. Often, this type of integration consists predominantly of digital circuitry, thus the efficiency of the digital in cost and power is key. MOS remains the most efficient integrated circuit technology for mixed-signal integration. The scaling of MOS process feature sizes combined with innovation in circuit design techniques have also opened new high bandwidth opportunities for MOS mixed- signal circuits. There is a trend to replace complex analog signal processing with digital signal pro- cessing (DSP). The use of over-sampled converters minimizes the requirements for pre- and post filtering. In spite of this trend, there are many applications where DSP is im- practical and where signal conditioning in the analog domain is necessary, particularly at high frequencies. Applications for high frequency filters are in data communications and the read channel of tape, disk and optical storage devices where filters bandlimit the data signal and perform amplitude and phase equalization. Other examples include filtering of radio and video signals. -
Lectures on the Local Semicircle Law for Wigner Matrices
Lectures on the local semicircle law for Wigner matrices Florent Benaych-Georges∗ Antti Knowlesy September 11, 2018 These notes provide an introduction to the local semicircle law from random matrix theory, as well as some of its applications. We focus on Wigner matrices, Hermitian random matrices with independent upper-triangular entries with zero expectation and constant variance. We state and prove the local semicircle law, which says that the eigenvalue distribution of a Wigner matrix is close to Wigner's semicircle distribution, down to spectral scales containing slightly more than one eigenvalue. This local semicircle law is formulated using the Green function, whose individual entries are controlled by large deviation bounds. We then discuss three applications of the local semicircle law: first, complete delocalization of the eigenvectors, stating that with high probability the eigenvectors are approximately flat; second, rigidity of the eigenvalues, giving large deviation bounds on the locations of the individual eigenvalues; third, a comparison argument for the local eigenvalue statistics in the bulk spectrum, showing that the local eigenvalue statistics of two Wigner matrices coincide provided the first four moments of their entries coincide. We also sketch further applications to eigenvalues near the spectral edge, and to the distribution of eigenvectors. arXiv:1601.04055v4 [math.PR] 10 Sep 2018 ∗Universit´eParis Descartes, MAP5. Email: [email protected]. yETH Z¨urich, Departement Mathematik. Email: [email protected]. -
Observational Astrophysics II January 18, 2007
Observational Astrophysics II JOHAN BLEEKER SRON Netherlands Institute for Space Research Astronomical Institute Utrecht FRANK VERBUNT Astronomical Institute Utrecht January 18, 2007 Contents 1RadiationFields 4 1.1 Stochastic processes: distribution functions, mean and variance . 4 1.2Autocorrelationandautocovariance........................ 5 1.3Wide-sensestationaryandergodicsignals..................... 6 1.4Powerspectraldensity................................ 7 1.5 Intrinsic stochastic nature of a radiation beam: Application of Bose-Einstein statistics ....................................... 10 1.5.1 Intermezzo: Bose-Einstein statistics . ................... 10 1.5.2 Intermezzo: Fermi Dirac statistics . ................... 13 1.6 Stochastic description of a radiation field in the thermal limit . 15 1.7 Stochastic description of a radiation field in the quantum limit . 20 2 Astronomical measuring process: information transfer 23 2.1Integralresponsefunctionforastronomicalmeasurements............ 23 2.2Timefiltering..................................... 24 2.2.1 Finiteexposureandtimeresolution.................... 24 2.2.2 Error assessment in sample average μT ................... 25 2.3 Data sampling in a bandlimited measuring system: Critical or Nyquist frequency 28 3 Indirect Imaging and Spectroscopy 31 3.1Coherence....................................... 31 3.1.1 The Visibility function . ................... 31 3.1.2 Young’sdualbeaminterferenceexperiment................ 31 3.1.3 Themutualcoherencefunction....................... 32 3.1.4 Interference -
The Equivalence of Digital and Analog Signal Processing *
Reprinted from I ~ FORMATIO:-; AXDCOXTROL, Volume 8, No. 5, October 1965 Copyright @ by AcademicPress Inc. Printed in U.S.A. INFORMATION A~D CO:";TROL 8, 455- 4G7 (19G5) The Equivalence of Digital and Analog Signal Processing * K . STEIGLITZ Departmentof Electrical Engineering, Princeton University, Princeton, l\rew J er.~e!J A specific isomorphism is constructed via the trallsform domaiIls between the analog signal spaceL2 (- 00, 00) and the digital signal space [2. It is then shown that the class of linear time-invariaIlt realizable filters is illvariallt under this isomorphism, thus demoIl- stratillg that the theories of processingsignals with such filters are identical in the digital and analog cases. This means that optimi - zatioll problems illvolving linear time-illvariant realizable filters and quadratic cost functions are equivalent in the discrete-time and the continuous-time cases, for both deterministic and random signals. FilIally , applicatiolls to the approximation problem for digital filters are discussed. LIST OF SY~IBOLS f (t ), g(t) continuous -time signals F (jCAJ), G(jCAJ) Fourier transforms of continuous -time signals A continuous -time filters , bounded linear transformations of I.J2(- 00, 00) {in}, {gn} discrete -time signals F (z), G (z) z-transforms of discrete -time signals A discrete -time filters , bounded linear transfornlations of ~ JL isomorphic mapping from L 2 ( - 00, 00) to l2 ~L 2 ( - 00, 00) space of Fourier transforms of functions in L 2( - 00,00) 512 space of z-transforms of sequences in l2 An(t ) nth Laguerre function * This work is part of a thesis submitted in partial fulfillment of requirements for the degree of Doctor of Engineering Scienceat New York University , and was supported partly by the National ScienceFoundation and partly by the Air Force Officeof Scientific Researchunder Contract No. -
Semi-Parametric Likelihood Functions for Bivariate Survival Data
Old Dominion University ODU Digital Commons Mathematics & Statistics Theses & Dissertations Mathematics & Statistics Summer 2010 Semi-Parametric Likelihood Functions for Bivariate Survival Data S. H. Sathish Indika Old Dominion University Follow this and additional works at: https://digitalcommons.odu.edu/mathstat_etds Part of the Mathematics Commons, and the Statistics and Probability Commons Recommended Citation Indika, S. H. S.. "Semi-Parametric Likelihood Functions for Bivariate Survival Data" (2010). Doctor of Philosophy (PhD), Dissertation, Mathematics & Statistics, Old Dominion University, DOI: 10.25777/ jgbf-4g75 https://digitalcommons.odu.edu/mathstat_etds/30 This Dissertation is brought to you for free and open access by the Mathematics & Statistics at ODU Digital Commons. It has been accepted for inclusion in Mathematics & Statistics Theses & Dissertations by an authorized administrator of ODU Digital Commons. For more information, please contact [email protected]. SEMI-PARAMETRIC LIKELIHOOD FUNCTIONS FOR BIVARIATE SURVIVAL DATA by S. H. Sathish Indika BS Mathematics, 1997, University of Colombo MS Mathematics-Statistics, 2002, New Mexico Institute of Mining and Technology MS Mathematical Sciences-Operations Research, 2004, Clemson University MS Computer Science, 2006, College of William and Mary A Dissertation Submitted to the Faculty of Old Dominion University in Partial Fulfillment of the Requirement for the Degree of DOCTOR OF PHILOSOPHY DEPARTMENT OF MATHEMATICS AND STATISTICS OLD DOMINION UNIVERSITY August 2010 Approved by: DayanandiN; Naik Larry D. Le ia M. Jones ABSTRACT SEMI-PARAMETRIC LIKELIHOOD FUNCTIONS FOR BIVARIATE SURVIVAL DATA S. H. Sathish Indika Old Dominion University, 2010 Director: Dr. Norou Diawara Because of the numerous applications, characterization of multivariate survival dis tributions is still a growing area of research. -
On the Duality of Regular and Local Functions
Preprints (www.preprints.org) | NOT PEER-REVIEWED | Posted: 24 May 2017 doi:10.20944/preprints201705.0175.v1 mathematics ISSN 2227-7390 www.mdpi.com/journal/mathematics Article On the Duality of Regular and Local Functions Jens V. Fischer Institute of Mathematics, Ludwig Maximilians University Munich, 80333 Munich, Germany; E-Mail: jens.fi[email protected]; Tel.: +49-8196-934918 1 Abstract: In this paper, we relate Poisson’s summation formula to Heisenberg’s uncertainty 2 principle. They both express Fourier dualities within the space of tempered distributions and 3 these dualities are furthermore the inverses of one another. While Poisson’s summation 4 formula expresses a duality between discretization and periodization, Heisenberg’s 5 uncertainty principle expresses a duality between regularization and localization. We define 6 regularization and localization on generalized functions and show that the Fourier transform 7 of regular functions are local functions and, vice versa, the Fourier transform of local 8 functions are regular functions. 9 Keywords: generalized functions; tempered distributions; regular functions; local functions; 10 regularization-localization duality; regularity; Heisenberg’s uncertainty principle 11 Classification: MSC 42B05, 46F10, 46F12 1 1.2 Introduction 13 Regularization is a popular trick in applied mathematics, see [1] for example. It is the technique 14 ”to approximate functions by more differentiable ones” [2]. Its terminology coincides moreover with 15 the terminology used in generalized function spaces. They contain two kinds of functions, ”regular 16 functions” and ”generalized functions”. While regular functions are being functions in the ordinary 17 functions sense which are infinitely differentiable in the ordinary functions sense, all other functions 18 become ”infinitely differentiable” in the ”generalized functions sense” [3]. -
Algorithmic Framework for X-Ray Nanocrystallographic Reconstruction in the Presence of the Indexing Ambiguity
Algorithmic framework for X-ray nanocrystallographic reconstruction in the presence of the indexing ambiguity Jeffrey J. Donatellia,b and James A. Sethiana,b,1 aDepartment of Mathematics and bLawrence Berkeley National Laboratory, University of California, Berkeley, CA 94720 Contributed by James A. Sethian, November 21, 2013 (sent for review September 23, 2013) X-ray nanocrystallography allows the structure of a macromolecule over multiple orientations. Although there has been some success in to be determined from a large ensemble of nanocrystals. How- determining structure from perfectly twinned data, reconstruction is ever, several parameters, including crystal sizes, orientations, and often infeasible without a good initial atomic model of the structure. incident photon flux densities, are initially unknown and images We present an algorithmic framework for X-ray nanocrystal- are highly corrupted with noise. Autoindexing techniques, com- lographic reconstruction which is based on directly reducing data monly used in conventional crystallography, can determine ori- variance and resolving the indexing ambiguity. First, we design entations using Bragg peak patterns, but only up to crystal lattice an autoindexing technique that uses both Bragg and non-Bragg symmetry. This limitation results in an ambiguity in the orienta- data to compute precise orientations, up to lattice symmetry. tions, known as the indexing ambiguity, when the diffraction Crystal sizes are then determined by performing a Fourier analysis pattern displays less symmetry than the lattice and leads to data around Bragg peak neighborhoods from a finely sampled low- that appear twinned if left unresolved. Furthermore, missing angle image, such as from a rear detector (Fig. 1). Next, we model phase information must be recovered to determine the imaged structure factor magnitudes for each reciprocal lattice point with object’s structure. -
Delta Functions and Distributions
When functions have no value(s): Delta functions and distributions Steven G. Johnson, MIT course 18.303 notes Created October 2010, updated March 8, 2017. Abstract x = 0. That is, one would like the function δ(x) = 0 for all x 6= 0, but with R δ(x)dx = 1 for any in- These notes give a brief introduction to the mo- tegration region that includes x = 0; this concept tivations, concepts, and properties of distributions, is called a “Dirac delta function” or simply a “delta which generalize the notion of functions f(x) to al- function.” δ(x) is usually the simplest right-hand- low derivatives of discontinuities, “delta” functions, side for which to solve differential equations, yielding and other nice things. This generalization is in- a Green’s function. It is also the simplest way to creasingly important the more you work with linear consider physical effects that are concentrated within PDEs, as we do in 18.303. For example, Green’s func- very small volumes or times, for which you don’t ac- tions are extremely cumbersome if one does not al- tually want to worry about the microscopic details low delta functions. Moreover, solving PDEs with in this volume—for example, think of the concepts of functions that are not classically differentiable is of a “point charge,” a “point mass,” a force plucking a great practical importance (e.g. a plucked string with string at “one point,” a “kick” that “suddenly” imparts a triangle shape is not twice differentiable, making some momentum to an object, and so on. -
Understanding the Lomb-Scargle Periodogram
Understanding the Lomb-Scargle Periodogram Jacob T. VanderPlas1 ABSTRACT The Lomb-Scargle periodogram is a well-known algorithm for detecting and char- acterizing periodic signals in unevenly-sampled data. This paper presents a conceptual introduction to the Lomb-Scargle periodogram and important practical considerations for its use. Rather than a rigorous mathematical treatment, the goal of this paper is to build intuition about what assumptions are implicit in the use of the Lomb-Scargle periodogram and related estimators of periodicity, so as to motivate important practical considerations required in its proper application and interpretation. Subject headings: methods: data analysis | methods: statistical 1. Introduction The Lomb-Scargle periodogram (Lomb 1976; Scargle 1982) is a well-known algorithm for de- tecting and characterizing periodicity in unevenly-sampled time-series, and has seen particularly wide use within the astronomy community. As an example of a typical application of this method, consider the data shown in Figure 1: this is an irregularly-sampled timeseries showing a single ob- ject from the LINEAR survey (Sesar et al. 2011; Palaversa et al. 2013), with un-filtered magnitude measured 280 times over the course of five and a half years. By eye, it is clear that the brightness of the object varies in time with a range spanning approximately 0.8 magnitudes, but what is not immediately clear is that this variation is periodic in time. The Lomb-Scargle periodogram is a method that allows efficient computation of a Fourier-like power spectrum estimator from such unevenly-sampled data, resulting in an intuitive means of determining the period of oscillation. -
Analog Signal Processing Circuits in Organic Transistor Technology a Dissertation Submitted to the Department of Electrical Engi
ANALOG SIGNAL PROCESSING CIRCUITS IN ORGANIC TRANSISTOR TECHNOLOGY A DISSERTATION SUBMITTED TO THE DEPARTMENT OF ELECTRICAL ENGINEERING AND THE COMMITTEE ON GRADUATE STUDIES OF STANFORD UNIVERSITY IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE DEGREE OF DOCTOR OF PHILOSOPHY Wei Xiong October 2010 © 2011 by Wei Xiong. All Rights Reserved. Re-distributed by Stanford University under license with the author. This work is licensed under a Creative Commons Attribution- Noncommercial 3.0 United States License. http://creativecommons.org/licenses/by-nc/3.0/us/ This dissertation is online at: http://purl.stanford.edu/fk938sr9136 ii I certify that I have read this dissertation and that, in my opinion, it is fully adequate in scope and quality as a dissertation for the degree of Doctor of Philosophy. Boris Murmann, Primary Adviser I certify that I have read this dissertation and that, in my opinion, it is fully adequate in scope and quality as a dissertation for the degree of Doctor of Philosophy. Zhenan Bao I certify that I have read this dissertation and that, in my opinion, it is fully adequate in scope and quality as a dissertation for the degree of Doctor of Philosophy. Robert Dutton Approved for the Stanford University Committee on Graduate Studies. Patricia J. Gumport, Vice Provost Graduate Education This signature page was generated electronically upon submission of this dissertation in electronic format. An original signed hard copy of the signature page is on file in University Archives. iii Abstract Low-voltage organic thin-film transistors offer potential for many novel applications. Because organic transistors can be fabricated near room temperature, they allow integrated circuits to be made on flexible plastic substrates. -
Fourier Ptychographic Wigner Distribution Deconvolution
FOURIER PTYCHOGRAPHIC WIGNER DISTRIBUTION DECONVOLUTION Justin Lee1, George Barbastathis2,3 1. Department of Health Sciences and Technology, Massachusetts Institute of Technology, 77 Massachusetts Avenue, Cambridge, MA 02139, USA 2. Department of Mechanical Engineering, Massachusetts Institute of Technology, 77 Massachusetts Avenue, Cambridge, MA 02139, USA 3. Singapore–MIT Alliance for Research and Technology (SMART) Centre Singapore 138602, Singapore Email: [email protected] KEY WORDS: Computational Imaging, phase retrieval, ptychography, Fourier ptychography Ptychography was first proposed as a method for phase retrieval by Hoppe in the late 1960s [1]. A typical ptychographic setup involves generation of illumination diversity via lateral translation of an object relative to its illumination (often called the probe) and measurement of the intensity of the far-field diffraction pattern/Fourier transform of the illuminated object. In the decades following Hoppes initial work, several important extensions to ptychography were developed, including an ability to simultaneously retrieve the amplitude and phase of both the probe and the object via application of an extended ptychographic iterative engine (ePIE) [2] and a non-iterative technique for phase retrieval using ptychographic information called Wigner Distribution Deconvolution (WDD) [3,4]. Most recently, the Fourier dual to ptychography was introduced by the authors in [5] as an experimentally simple technique for phase retrieval using an LED array as the illumination source in a standard microscope setup. In Fourier ptychography, lateral translation of the object relative to the probe occurs in frequency space and is obtained by tilting the angle of illumination (while holding the pupil function of the imaging system constant). In the past few years, several improvements to Fourier ptychography have been introduced, including an ePIE-like technique for simultaneous pupil function and object retrieval [6].