Sampling—50 Years After Shannon
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Sampling Signals on Graphs from Theory to Applications
1 Sampling Signals on Graphs From Theory to Applications Yuichi Tanaka, Yonina C. Eldar, Antonio Ortega, and Gene Cheung Abstract The study of sampling signals on graphs, with the goal of building an analog of sampling for standard signals in the time and spatial domains, has attracted considerable attention recently. Beyond adding to the growing theory on graph signal processing (GSP), sampling on graphs has various promising applications. In this article, we review current progress on sampling over graphs focusing on theory and potential applications. Although most methodologies used in graph signal sampling are designed to parallel those used in sampling for standard signals, sampling theory for graph signals significantly differs from the theory of Shannon–Nyquist and shift-invariant sampling. This is due in part to the fact that the definitions of several important properties, such as shift invariance and bandlimitedness, are different in GSP systems. Throughout this review, we discuss similarities and differences between standard and graph signal sampling and highlight open problems and challenges. I. INTRODUCTION Sampling is one of the fundamental tenets of digital signal processing (see [1] and references therein). As such, it has been studied extensively for decades and continues to draw considerable research efforts. Standard sampling theory relies on concepts of frequency domain analysis, shift invariant (SI) signals, and bandlimitedness [1]. Sampling of time and spatial domain signals in shift-invariant spaces is one of the most important building blocks of digital signal processing systems. However, in the big data era, the signals we need to process often have other types of connections and structure, such as network signals described by graphs. -
Lecture19.Pptx
Outline Foundations of Computer Graphics (Fall 2012) §. Basic ideas of sampling, reconstruction, aliasing CS 184, Lectures 19: Sampling and Reconstruction §. Signal processing and Fourier analysis §. http://inst.eecs.berkeley.edu /~cs184 Implementation of digital filters §. Section 14.10 of FvDFH (you really should read) §. Post-raytracing lectures more advanced topics §. No programming assignment §. But can be tested (at high level) in final Acknowledgements: Thomas Funkhouser and Pat Hanrahan Some slides courtesy Tom Funkhouser Sampling and Reconstruction Sampling and Reconstruction §. An image is a 2D array of samples §. Discrete samples from real-world continuous signal (Spatial) Aliasing (Spatial) Aliasing §. Jaggies probably biggest aliasing problem 1 Sampling and Aliasing Image Processing pipeline §. Artifacts due to undersampling or poor reconstruction §. Formally, high frequencies masquerading as low §. E.g. high frequency line as low freq jaggies Outline Motivation §. Basic ideas of sampling, reconstruction, aliasing §. Formal analysis of sampling and reconstruction §. Signal processing and Fourier analysis §. Important theory (signal-processing) for graphics §. Implementation of digital filters §. Also relevant in rendering, modeling, animation §. Section 14.10 of FvDFH Ideas Sampling Theory Analysis in the frequency (not spatial) domain §. Signal (function of time generally, here of space) §. Sum of sine waves, with possibly different offsets (phase) §. Each wave different frequency, amplitude §. Continuous: defined at all points; discrete: on a grid §. High frequency: rapid variation; Low Freq: slow variation §. Images are converting continuous to discrete. Do this sampling as best as possible. §. Signal processing theory tells us how best to do this §. Based on concept of frequency domain Fourier analysis 2 Fourier Transform Fourier Transform §. Tool for converting from spatial to frequency domain §. -
And Bandlimiting IMAHA, November 14, 2009, 11:00–11:40
Time- and bandlimiting IMAHA, November 14, 2009, 11:00{11:40 Joe Lakey (w Scott Izu and Jeff Hogan)1 November 16, 2009 Joe Lakey (w Scott Izu and Jeff Hogan) Time- and bandlimiting Joe Lakey (w Scott Izu and Jeff Hogan) Time- and bandlimiting Joe Lakey (w Scott Izu and Jeff Hogan) Time- and bandlimiting Joe Lakey (w Scott Izu and Jeff Hogan) Time- and bandlimiting sampling theory and history Time- and bandlimiting, history, definitions and basic properties connecting sampling and time- and bandlimiting The multiband case Outline Joe Lakey (w Scott Izu and Jeff Hogan) Time- and bandlimiting Time- and bandlimiting, history, definitions and basic properties connecting sampling and time- and bandlimiting The multiband case Outline sampling theory and history Joe Lakey (w Scott Izu and Jeff Hogan) Time- and bandlimiting connecting sampling and time- and bandlimiting The multiband case Outline sampling theory and history Time- and bandlimiting, history, definitions and basic properties Joe Lakey (w Scott Izu and Jeff Hogan) Time- and bandlimiting The multiband case Outline sampling theory and history Time- and bandlimiting, history, definitions and basic properties connecting sampling and time- and bandlimiting Joe Lakey (w Scott Izu and Jeff Hogan) Time- and bandlimiting Outline sampling theory and history Time- and bandlimiting, history, definitions and basic properties connecting sampling and time- and bandlimiting The multiband case Joe Lakey (w Scott Izu and Jeff Hogan) Time- and bandlimiting R −2πitξ Fourier transform: bf (ξ) = f (t) e dt R _ Bandlimiting: -
An Overview of Wavelet Transform Concepts and Applications
An overview of wavelet transform concepts and applications Christopher Liner, University of Houston February 26, 2010 Abstract The continuous wavelet transform utilizing a complex Morlet analyzing wavelet has a close connection to the Fourier transform and is a powerful analysis tool for decomposing broadband wavefield data. A wide range of seismic wavelet applications have been reported over the last three decades, and the free Seismic Unix processing system now contains a code (succwt) based on the work reported here. Introduction The continuous wavelet transform (CWT) is one method of investigating the time-frequency details of data whose spectral content varies with time (non-stationary time series). Moti- vation for the CWT can be found in Goupillaud et al. [12], along with a discussion of its relationship to the Fourier and Gabor transforms. As a brief overview, we note that French geophysicist J. Morlet worked with non- stationary time series in the late 1970's to find an alternative to the short-time Fourier transform (STFT). The STFT was known to have poor localization in both time and fre- quency, although it was a first step beyond the standard Fourier transform in the analysis of such data. Morlet's original wavelet transform idea was developed in collaboration with the- oretical physicist A. Grossmann, whose contributions included an exact inversion formula. A series of fundamental papers flowed from this collaboration [16, 12, 13], and connections were soon recognized between Morlet's wavelet transform and earlier methods, including harmonic analysis, scale-space representations, and conjugated quadrature filters. For fur- ther details, the interested reader is referred to Daubechies' [7] account of the early history of the wavelet transform. -
A Low Bit Rate Audio Codec Using Wavelet Transform
Navpreet Singh, Mandeep Kaur,Rajveer Kaur / International Journal of Engineering Research and Applications (IJERA) ISSN: 2248-9622 www.ijera.com Vol. 3, Issue 4, Jul-Aug 2013, pp.2222-2228 An Enhanced Low Bit Rate Audio Codec Using Discrete Wavelet Transform Navpreet Singh1, Mandeep Kaur2, Rajveer Kaur3 1,2(M. tech Students, Department of ECE, Guru kashi University, Talwandi Sabo(BTI.), Punjab, INDIA 3(Asst. Prof. Department of ECE, Guru Kashi University, Talwandi Sabo (BTI.), Punjab,INDIA Abstract Audio coding is the technology to represent information and perceptually irrelevant signal audio in digital form with as few bits as possible components can be separated and later removed. This while maintaining the intelligibility and quality class includes techniques such as subband coding; required for particular application. Interest in audio transform coding, critical band analysis, and masking coding is motivated by the evolution to digital effects. The second class takes advantage of the communications and the requirement to minimize statistical redundancy in audio signal and applies bit rate, and hence conserve bandwidth. There is some form of digital encoding. Examples of this class always a tradeoff between lowering the bit rate and include entropy coding in lossless compression and maintaining the delivered audio quality and scalar/vector quantization in lossy compression [1]. intelligibility. The wavelet transform has proven to Digital audio compression allows the be a valuable tool in many application areas for efficient storage and transmission of audio data. The analysis of nonstationary signals such as image and various audio compression techniques offer different audio signals. In this paper a low bit rate audio levels of complexity, compressed audio quality, and codec algorithm using wavelet transform has been amount of data compression. -
Image Compression Techniques by Using Wavelet Transform
View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by International Institute for Science, Technology and Education (IISTE): E-Journals Journal of Information Engineering and Applications www.iiste.org ISSN 2224-5782 (print) ISSN 2225-0506 (online) Vol 2, No.5, 2012 Image Compression Techniques by using Wavelet Transform V. V. Sunil Kumar 1* M. Indra Sena Reddy 2 1. Dept. of CSE, PBR Visvodaya Institute of Tech & Science, Kavali, Nellore (Dt), AP, INDIA 2. School of Computer science & Enginering, RGM College of Engineering & Tech. Nandyal, A.P, India * E-mail of the corresponding author: [email protected] Abstract This paper is concerned with a certain type of compression techniques by using wavelet transforms. Wavelets are used to characterize a complex pattern as a series of simple patterns and coefficients that, when multiplied and summed, reproduce the original pattern. The data compression schemes can be divided into lossless and lossy compression. Lossy compression generally provides much higher compression than lossless compression. Wavelets are a class of functions used to localize a given signal in both space and scaling domains. A MinImage was originally created to test one type of wavelet and the additional functionality was added to Image to support other wavelet types, and the EZW coding algorithm was implemented to achieve better compression. Keywords: Wavelet Transforms, Image Compression, Lossless Compression, Lossy Compression 1. Introduction Digital images are widely used in computer applications. Uncompressed digital images require considerable storage capacity and transmission bandwidth. Efficient image compression solutions are becoming more critical with the recent growth of data intensive, multimedia based web applications. -
JPEG2000: Wavelets in Image Compression
EE678 WAVELETS APPLICATION ASSIGNMENT 1 JPEG2000: Wavelets In Image Compression Group Members: Qutubuddin Saifee [email protected] 01d07009 Ankur Gupta [email protected] 01d070013 Nishant Singh [email protected] 01d07019 Abstract During the past decade, with the birth of wavelet theory and multiresolution analysis, image processing techniques based on wavelet transform have been extensively studied and tremendously improved. JPEG 2000 uses wavelet transform and provides an integrated toolbox to better address increasing needs for compression. In this report, we study the basic concepts of JPEG2000, the LeGall 5/3 and Daubechies 9/7 wavelets used in it and finally Embedded zerotree wavelet coding and Set Partitioning in Hierarchical Trees. Index Terms Wavelets, JPEG2000, Image Compression, LeGall, Daubechies, EZW, SPIHT. I. INTRODUCTION I NCE the mid 1980s, members from both the International Telecommunications Union (ITU) and the International SOrganization for Standardization (ISO) have been working together to establish a joint international standard for the compression of grayscale and color still images. This effort has been known as JPEG, the Joint Photographic Experts Group. The process was such that, after evaluating a number of coding schemes, the JPEG members selected a discrete cosine transform( DCT)-based method in 1988. From 1988 to 1990, the JPEG group continued its work by simulating, testing and documenting the algorithm. JPEG became Draft International Standard (DIS) in 1991 and International Standard (IS) in 1992. With the continual expansion of multimedia and Internet applications, the needs and requirements of the technologies used grew and evolved. In March 1997, a new call for contributions was launched for the development of a new standard for the compression of still images, the JPEG2000 standard. -
Information Theory
Information Theory Professor John Daugman University of Cambridge Computer Science Tripos, Part II Michaelmas Term 2016/17 H(X,Y) I(X;Y) H(X|Y) H(Y|X) H(X) H(Y) 1 / 149 Outline of Lectures 1. Foundations: probability, uncertainty, information. 2. Entropies defined, and why they are measures of information. 3. Source coding theorem; prefix, variable-, and fixed-length codes. 4. Discrete channel properties, noise, and channel capacity. 5. Spectral properties of continuous-time signals and channels. 6. Continuous information; density; noisy channel coding theorem. 7. Signal coding and transmission schemes using Fourier theorems. 8. The quantised degrees-of-freedom in a continuous signal. 9. Gabor-Heisenberg-Weyl uncertainty relation. Optimal \Logons". 10. Data compression codes and protocols. 11. Kolmogorov complexity. Minimal description length. 12. Applications of information theory in other sciences. Reference book (*) Cover, T. & Thomas, J. Elements of Information Theory (second edition). Wiley-Interscience, 2006 2 / 149 Overview: what is information theory? Key idea: The movements and transformations of information, just like those of a fluid, are constrained by mathematical and physical laws. These laws have deep connections with: I probability theory, statistics, and combinatorics I thermodynamics (statistical physics) I spectral analysis, Fourier (and other) transforms I sampling theory, prediction, estimation theory I electrical engineering (bandwidth; signal-to-noise ratio) I complexity theory (minimal description length) I signal processing, representation, compressibility As such, information theory addresses and answers the two fundamental questions which limit all data encoding and communication systems: 1. What is the ultimate data compression? (answer: the entropy of the data, H, is its compression limit.) 2. -
Lecture 16 Bandlimiting and Nyquist Criterion
ELG3175 Introduction to Communication Systems Lecture 16 Bandlimiting and Nyquist Criterion Bandlimiting and ISI • Real systems are usually bandlimited. • When a signal is bandlimited in the frequency domain, it is usually smeared in the time domain. This smearing results in intersymbol interference (ISI). • The only way to avoid ISI is to satisfy the 1st Nyquist criterion. • For an impulse response this means at sampling instants having only one nonzero sample. Lecture 12 Band-limited Channels and Intersymbol Interference • Rectangular pulses are suitable for infinite-bandwidth channels (practically – wideband). • Practical channels are band-limited -> pulses spread in time and are smeared into adjacent slots. This is intersymbol interference (ISI). Input binary waveform Individual pulse response Received waveform Eye Diagram • Convenient way to observe the effect of ISI and channel noise on an oscilloscope. Eye Diagram • Oscilloscope presentations of a signal with multiple sweeps (triggered by a clock signal!), each is slightly larger than symbol interval. • Quality of a received signal may be estimated. • Normal operating conditions (no ISI, no noise) -> eye is open. • Large ISI or noise -> eye is closed. • Timing error allowed – width of the eye, called eye opening (preferred sampling time – at the largest vertical eye opening). • Sensitivity to timing error -> slope of the open eye evaluated at the zero crossing point. • Noise margin -> the height of the eye opening. Pulse shapes and bandwidth • For PAM: L L sPAM (t) = ∑ai p(t − iTs ) = p(t)*∑aiδ (t − iTs ) i=0 i=0 L Let ∑aiδ (t − iTs ) = y(t) i=0 Then SPAM ( f ) = P( f )Y( f ) BPAM = Bp. -
Frames and Other Bases in Abstract and Function Spaces Novel Methods in Harmonic Analysis, Volume 1
Applied and Numerical Harmonic Analysis Isaac Pesenson Quoc Thong Le Gia Azita Mayeli Hrushikesh Mhaskar Ding-Xuan Zhou Editors Frames and Other Bases in Abstract and Function Spaces Novel Methods in Harmonic Analysis, Volume 1 Applied and Numerical Harmonic Analysis Series Editor John J. Benedetto University of Maryland College Park, MD, USA Editorial Advisory Board Akram Aldroubi Gitta Kutyniok Vanderbilt University Technische Universität Berlin Nashville, TN, USA Berlin, Germany Douglas Cochran Mauro Maggioni Arizona State University Duke University Phoenix, AZ, USA Durham, NC, USA Hans G. Feichtinger Zuowei Shen University of Vienna National University of Singapore Vienna, Austria Singapore, Singapore Christopher Heil Thomas Strohmer Georgia Institute of Technology University of California Atlanta, GA, USA Davis, CA, USA Stéphane Jaffard Yang Wang University of Paris XII Michigan State University Paris, France East Lansing, MI, USA Jelena Kovaceviˇ c´ Carnegie Mellon University Pittsburgh, PA, USA More information about this series at http://www.springer.com/series/4968 Isaac Pesenson • Quoc Thong Le Gia Azita Mayeli • Hrushikesh Mhaskar Ding-Xuan Zhou Editors Frames and Other Bases in Abstract and Function Spaces Novel Methods in Harmonic Analysis, Volume 1 Editors Isaac Pesenson Quoc Thong Le Gia Department of Mathematics School of Mathematics and Statistics Temple University University of New South Wales Philadelphia, PA, USA Sydney, NSW, Australia Azita Mayeli Hrushikesh Mhaskar Department of Mathematics Institute of Mathematical -
Comparison of Image Compressions: Analog Transformations P
Proceedings Comparison of Image Compressions: Analog † Transformations P Jose Balsa P CITIC Research Center, Universidade da Coruña (University of A Coruña), 15071 A Coruña, Spain; [email protected] † Presented at the 3rd XoveTIC Conference, A Coruña, Spain, 8–9 October 2020. Published: 21 August 2020 Abstract: A comparison between the four most used transforms, the discrete Fourier transform (DFT), discrete cosine transform (DCT), the Walsh–Hadamard transform (WHT) and the Haar- wavelet transform (DWT), for the transmission of analog images, varying their compression and comparing their quality, is presented. Additionally, performance tests are done for different levels of white Gaussian additive noise. Keywords: analog image transformation; analog image compression; analog image quality 1. Introduction Digitized image coding systems employ reversible mathematical transformations. These transformations change values and function domains in order to rearrange information in a way that condenses information important to human vision [1]. In the new domain, it is possible to filter out relevant information and discard information that is irrelevant or of lesser importance for image quality [2]. Both digital and analog systems use the same transformations in source coding. Some examples of digital systems that employ these transformations are JPEG, M-JPEG, JPEG2000, MPEG- 1, 2, 3 and 4, DV and HDV, among others. Although digital systems after transformation and filtering make use of digital lossless compression techniques, such as Huffman. In this work, we aim to make a comparison of the most commonly used transformations in state- of-the-art image compression systems. Typically, the transformations used to compress analog images work either on the entire image or on regions of the image. -
Ideal C/D Converter
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.341: Discrete-Time Signal Processing OpenCourseWare 2006 Lecture 4 DT Processing of CT Signals & CT Processing of DT Signals: Fractional Delay Reading: Sections 4.1 - 4.5 in Oppenheim, Schafer & Buck (OSB). A typical discrete-time system used to process a continuous-time signal is shown in OSB Figure 4.15. T is the sampling/reconstruction period for the C/D and the D/C converters. When the input signal is bandlimited, the effective system in Figure 4.15 (b) is equivalent to the continuous-time system shown in Figure 4.15 (a). Ideal C/D Converter The relationship between the input and output signals in an ideal C/D converter as depicted in OSB Figure 4.1 is: Time Domain: x[n] = xc(nT ) j! 1 P1 ! 2¼k Frequency Domain: X(e ) = T k=¡1 Xc(j( T ¡ T )); j! where X(e ) and Xc(jΩ) are the DTFT and CTFT of x[n], xc(t). These relationships are illustrated in the figures below. See Section 4.2 of OSB for their derivations and more detailed descriptions of C/D converters. Time/Frequency domain representation of C/D conversion 1 In the frequency domain, a C/D converter can be thought of in three stages: periodic 1 repetition of the spectrum at 2¼, normalization of frequency by , and scaling of the amplitude T 1 ! by . The CT frequency Ω is related to the DT frequency ! by Ω = . If x (t) is not T T c s s bandlimited to ¡ 2 < Ωmax < 2 , repetition at 2¼ will cause aliasing.