Speech Compression Using Discrete Wavelet Transform and Discrete Cosine Transform

Total Page:16

File Type:pdf, Size:1020Kb

Speech Compression Using Discrete Wavelet Transform and Discrete Cosine Transform International Journal of Engineering Research & Technology (IJERT) ISSN: 2278-0181 Vol. 1 Issue 5, July - 2012 Speech Compression Using Discrete Wavelet Transform and Discrete Cosine Transform Smita Vatsa, Dr. O. P. Sahu M. Tech (ECE Student ) Professor Department of Electronicsand Department of Electronics and Communication Engineering Communication Engineering NIT Kurukshetra, India NIT Kurukshetra, India Abstract reproduced with desired level of quality. Main approaches of speech compression used today are Aim of this paper is to explain and implement waveform coding, transform coding and parametric transform based speech compression techniques. coding. Waveform coding attempts to reproduce Transform coding is based on compressing signal input signal waveform at the output. In transform by removing redundancies present in it. Speech coding at the beginning of procedure signal is compression (coding) is a technique to transform transformed into frequency domain, afterwards speech signal into compact format such that speech only dominant spectral features of signal are signal can be transmitted and stored with reduced maintained. In parametric coding signals are bandwidth and storage space respectively represented through a small set of parameters that .Objective of speech compression is to enhance can describe it accurately. Parametric coders transmission and storage capacity. In this paper attempts to produce a signal that sounds like Discrete wavelet transform and Discrete cosine original speechwhether or not time waveform transform based speech compression techniques resembles the original. are implemented with Run length encoding, By removing redundancy between Huffman encoding and Run length encoding neighbouring samples signal can be compressed. In followed by Huffman encoding. Reconstructed this paper we have implemented compression speech obtained from implementation are technique in two step, in 1st step a transform compared on the basis of Compression Factor(CF), function is applied on speech signal to get result Signal to noise ratio (SNR), Peak signal to noise with a new set of data with smaller values and more ratio (PSNR), Normalized root mean square error repetition, 2nd step is coding(compression) step, this (NRMSE), Retained signal energy (RSE). Results step willrepresent the data set in its minimal form shows that Discrete wavelet transform with RUN by using encoding techniques such as Run Length length encoding followed by Huffman encoding encoding, Huffman encoding, Run lengthencoding gives higher compression with intelligible followed by Huffman encoding. Performance reconstructed speech. measurescompression factor(CF),signal to noise ratio(SNR),peak signal to noise 1.Introduction ratio(PSNR),normalized root mean square error(NRMSE), retained signal energy (RSE) is Objective of speech is communication measured for reconstructed speech obtained from whether face to face or cell phone to cell phone. A DWT and DCT based speech compression huge amount of data is a big issue for transmission techniques. A comparative analysis of performance or storage. Speech compression is the technology of transform technique based speech compression of converting human speech into an efficiently systems for 3 different encoding methods is done. encoded representation that can later be decoded to In this paper a brief introduction of speech produce a close approximation of the original compression technique is presented in section 2, signal. Major objective of speech compression is to section 2 discusses Transform techniques, in represent speech with less or few number of bits section 3 implementation of speech compression with level of quality. Speech coding means based different transform technique with several representing speech signal in an efficient way for encoding methods is explained, section 4 presents transmission and storage such a way it can be simulation results obtained from MATLAB 7.12.0, followed by concluding remarks in section 5. www.ijert.org 1 International Journal of Engineering Research & Technology (IJERT) ISSN: 2278-0181 Vol. 1 Issue 5, July - 2012 2. Overview of transform techniques 2.2 Discrete cosine transform In this paper DWT and DCT transforms are We can use DCT for speech compression exploited for speech compression. because of high correlation in adjacent coefficient. Energy compaction property of DCT is good, we 2.1. Discrete wavelet transform can often reconstruct a sequence very accurately from only a few DCT coefficients, this property of A wavelet is a waveform of effectively limited DCT is very useful for applications requiring data duration that has an average value zero. reduction[7]. Fundamental idea behind wavelet is to analyse Discrete cosine transform of 1-D sequence x(n) of according to scale[4]. Wavelets are functions that length N is satisfy certain mathematical requirements and are 1/2 2 N1 (2nm 1) used in representing data or other functions.wavelet X( m ) c x ( n )cos functions are localized in space. In contrast Fourier m NN m0 2 sine and cosine functions are non-local and are active for all time t. This localization feature, along Where m=0,1,…..,N-1. with wavelets localization of frequency, makes The inverse discrete cosine transform is defined as many functions and operators using wavelets 1/2 N1 “sparse”, when transformed into the wavelet 2 (2nm 1) x( n ) c X ( m )cos domain. This sparseness, in turn results in a number m NN m0 2 of useful applications such as data compression, detecting features in images and de -noising In both equations c is defined as signals. The Discrete Wavelet Transform (DWT) m C = (1/2)1/2 for m=0 involves choosing scales and positions based on m 1 for m≠0 powers of two , so called dyadic scales and positions. The mother wavelet is rescaled or dilated, by powers of two and translated by 3. METHODOLOGY FOR COMPRESSION OF integers.Specifically, a function f(t) ∈ L2(R) (defines space of square integrable functions) can SPEECH SIGNAL be represented as[9]. In this paper we are implementing speech compression technique based on transform method L jLi.e. DCT and DWT. When we apply wavelet ft() djk (,)(2 tk ) aLK (,)(2 tk transform ) technique on speech signal original j1 k k Signal can be represented in terms of a wavelet The function ψ(t) is known as the mother wavelet, expansion (using coefficients in a linear while φ(t) is known as the scaling function. The set combination of wavelet function),similarly in case of functions of DCT transform speech can be represented in terms of DCT coefficient. Thus data operation can 2L (2 Lt k );2 j (2 j t k )| j L ,,, j k L Zbe performed using just the corresponding DWT and DCT coefficients. Transform techniques and where Z is the set of integers, is an orthonormal thresholding does not actually compress a signal, it basis for L2(R). simply provides information about the signal, The numbers a(L,k) are known as the which allows the data to be compressed by standard approximation coefficients at scale L, while d(j,k) encoding techniques. are known as the detail coefficients at scale j. Speech compression is achieved by neglecting The approximation and detail coefficients can be small coefficients as insignificant data and expressed as: discarding them and then applying quantization 1 and encoding scheme on coefficients. Speech a( L , k ) f ( t ) (2L t k ) dt compression algorithm is performed in following L 2 steps: (1). Transform technique on speech signal 1 j (2). Thresholding of transformed coefficient. d( j , k ) f ( t ) (2 t k ) dt 2 j (3). Quantization (4). Encoding wavelets with a high number of vanishing By following above 4 steps we will get a moments lead to a more compact signal compressed speech signal. For reconstruction of representation and are hence useful in coding speech signal inverse of above processes are applications[9]. performed i.e. decoding, de-quantization, inverse www.ijert.org 2 International Journal of Engineering Research & Technology (IJERT) ISSN: 2278-0181 Vol. 1 Issue 5, July - 2012 transform. 4 different speech signal samples are selected. Step size ∆ is obtained using above 3 used for data base. Data base used are in .wav parameters. format, sampled at 8khz. ∆=(mmax-mmin)/L Explanation for compression algorithm: Input is divided into L+1 levels with equal interval 3.1. Transform technique on speech signal size ranging from m to m to create the min max DCT and DWT transform methods are used on quantization table. Coefficient values are quantized speech signal. We can reconstruct a sequence very to integer values[1]. accurately from only a few DCT coefficient, this property of DCT transform is used for data compression. Localization feature of wavelet along with time-frequency resolution property makes Speech signal them well suited for speech coding. Sparse coding of wavelet is utilized for compression application. The idea behind signal compression using wavelets is primarily linked to the relative scarceness of the Transform Reconstructed speech signal wavelet domain representation for the signal. techniques Wavelets concentrate speech information (energy and perception) into a few neighbouring Thresholding of Inverse transform coefficients [8]. coefficients Therefore as a result of taking the wavelet transform of a signal, many coefficients will either be zero or have negligible magnitudes. Data Quantization De-quantization compression is then achieved by treating small valued coefficients as insignificant data and thus discarding them. Compressed Encoding Decoding data 3.2. Thresholding After getting coefficient signal from different Fig. 1. Block diagram of compression technique transform methods thresholding is applied to signal. In case of DCT as we mentioned earlier that 3.4. Encoding very few DCT coefficient represent 99% of signal In this paper we are using 3 different approaches energy, hence threshold based on above mentioned property is calculated and applied to coefficient for encoding of quantized coefficients: obtained from DCT transform, coefficients of which values are less than thresholds value are 1) Run length encoding: Coefficients obtained truncated ie.
Recommended publications
  • Speech Coding and Compression
    Introduction to voice coding/compression PCM coding Speech vocoders based on speech production LPC based vocoders Speech over packet networks Speech coding and compression Corso di Networked Multimedia Systems Master Universitario di Primo Livello in Progettazione e Gestione di Sistemi di Rete Carlo Drioli Università degli Studi di Verona Facoltà di Scienze Matematiche, Dipartimento di Informatica Fisiche e Naturali Speech coding and compression Introduction to voice coding/compression PCM coding Speech vocoders based on speech production LPC based vocoders Speech over packet networks Speech coding and compression: OUTLINE Introduction to voice coding/compression PCM coding Speech vocoders based on speech production LPC based vocoders Speech over packet networks Speech coding and compression Introduction to voice coding/compression PCM coding Speech vocoders based on speech production LPC based vocoders Speech over packet networks Introduction Approaches to voice coding/compression I Waveform coders (PCM) I Voice coders (vocoders) Quality assessment I Intelligibility I Naturalness (involves speaker identity preservation, emotion) I Subjective assessment: Listening test, Mean Opinion Score (MOS), Diagnostic acceptability measure (DAM), Diagnostic Rhyme Test (DRT) I Objective assessment: Signal to Noise Ratio (SNR), spectral distance measures, acoustic cues comparison Speech coding and compression Introduction to voice coding/compression PCM coding Speech vocoders based on speech production LPC based vocoders Speech over packet networks
    [Show full text]
  • Digital Speech Processing— Lecture 17
    Digital Speech Processing— Lecture 17 Speech Coding Methods Based on Speech Models 1 Waveform Coding versus Block Processing • Waveform coding – sample-by-sample matching of waveforms – coding quality measured using SNR • Source modeling (block processing) – block processing of signal => vector of outputs every block – overlapped blocks Block 1 Block 2 Block 3 2 Model-Based Speech Coding • we’ve carried waveform coding based on optimizing and maximizing SNR about as far as possible – achieved bit rate reductions on the order of 4:1 (i.e., from 128 Kbps PCM to 32 Kbps ADPCM) at the same time achieving toll quality SNR for telephone-bandwidth speech • to lower bit rate further without reducing speech quality, we need to exploit features of the speech production model, including: – source modeling – spectrum modeling – use of codebook methods for coding efficiency • we also need a new way of comparing performance of different waveform and model-based coding methods – an objective measure, like SNR, isn’t an appropriate measure for model- based coders since they operate on blocks of speech and don’t follow the waveform on a sample-by-sample basis – new subjective measures need to be used that measure user-perceived quality, intelligibility, and robustness to multiple factors 3 Topics Covered in this Lecture • Enhancements for ADPCM Coders – pitch prediction – noise shaping • Analysis-by-Synthesis Speech Coders – multipulse linear prediction coder (MPLPC) – code-excited linear prediction (CELP) • Open-Loop Speech Coders – two-state excitation
    [Show full text]
  • Speech Coding Using Code Excited Linear Prediction
    ISSN : 0976-8491 (Online) | ISSN : 2229-4333 (Print) IJCST VOL . 5, Iss UE SPL - 2, JAN - MAR C H 2014 Speech Coding Using Code Excited Linear Prediction 1Hemanta Kumar Palo, 2Kailash Rout 1ITER, Siksha ‘O’ Anusandhan University, Bhubaneswar, Odisha, India 2Gandhi Institute For Technology, Bhubaneswar, Odisha, India Abstract The main problem with the speech coding system is the optimum utilization of channel bandwidth. Due to this the speech signal is coded by using as few bits as possible to get low bit-rate speech coders. As the bit rate of the coder goes low, the intelligibility, SNR and overall quality of the speech signal decreases. Hence a comparative analysis is done of two different types of speech coders in this paper for understanding the utility of these coders in various applications so as to reduce the bandwidth and by different speech coding techniques and by reducing the number of bits without any appreciable compromise on the quality of speech. Hindi language has different number of stops than English , hence the performance of the coders must be checked on different languages. The main objective of this paper is to develop speech coders capable of producing high quality speech at low data rates. The focus of this paper is the development and testing of voice coding systems which cater for the above needs. Keywords PCM, DPCM, ADPCM, LPC, CELP Fig. 1: Speech Production Process I. Introduction Speech coding or speech compression is the compact digital The two types of speech sounds are voiced and unvoiced [1]. representations of voice signals [1-3] for the purpose of efficient They produce different sounds and spectra due to their differences storage and transmission.
    [Show full text]
  • An Ultra Low-Power Miniature Speech CODEC at 8 Kb/S and 16 Kb/S Robert Brennan, David Coode, Dustin Griesdorf, Todd Schneider Dspfactory Ltd
    To appear in ICSPAT 2000 Proceedings, October 16-19, 2000, Dallas, TX. An Ultra Low-power Miniature Speech CODEC at 8 kb/s and 16 kb/s Robert Brennan, David Coode, Dustin Griesdorf, Todd Schneider Dspfactory Ltd. 611 Kumpf Drive, Unit 200 Waterloo, Ontario, N2V 1K8 Canada Abstract SmartCODEC platform consists of an effi- cient, block-floating point, oversampled This paper describes a CODEC implementa- Weighted OverLap-Add (WOLA) filterbank, a tion on an ultra low-power miniature applica- software-programmable dual-Harvard 16-bit tion specific signal processor (ASSP) designed DSP core, two high fidelity 14-bit A/D con- for mobile audio signal processing applica- verters, a 14-bit D/A converter and a flexible tions. The CODEC records speech to and set of peripherals [1]. The system hardware plays back from a serial flash memory at data architecture (Figure 1) was designed to enable rates of 16 and 8 kb/s, with a bandwidth of 4 memory upgrades. Removable memory cards kHz. This CODEC consumes only 1 mW in a or more power-efficient memory could be package small enough for use in a range of substituted for the serial flash memory. demanding portable applications. Results, improvements and applications are also dis- The CODEC communicates with the flash cussed. memory over an integrated SPI port that can transfer data at rates up to 80 kb/s. For this application, the port is configured to block 1. Introduction transfer frame packets every 14 ms. e Speech coding is ubiquitous. Increasing de- c i WOLA Filterbank v mands for portability and fidelity coupled with A/D e and D the desire for reduced storage and bandwidth o i l Programmable d o r u utilization have increased the demand for and t D/A DSP Core A deployment of speech CODECs.
    [Show full text]
  • Cognitive Speech Coding Milos Cernak, Senior Member, IEEE, Afsaneh Asaei, Senior Member, IEEE, Alexandre Hyafil
    1 Cognitive Speech Coding Milos Cernak, Senior Member, IEEE, Afsaneh Asaei, Senior Member, IEEE, Alexandre Hyafil Abstract—Speech coding is a field where compression ear and undergoes a highly complex transformation paradigms have not changed in the last 30 years. The before it is encoded efficiently by spikes at the auditory speech signals are most commonly encoded with com- nerve. This great efficiency in information representation pression methods that have roots in Linear Predictive has inspired speech engineers to incorporate aspects of theory dating back to the early 1940s. This paper tries to cognitive processing in when developing efficient speech bridge this influential theory with recent cognitive studies applicable in speech communication engineering. technologies. This tutorial article reviews the mechanisms of speech Speech coding is a field where research has slowed perception that lead to perceptual speech coding. Then considerably in recent years. This has occurred not it focuses on human speech communication and machine because it has achieved the ultimate in minimizing bit learning, and application of cognitive speech processing in rate for transparent speech quality, but because recent speech compression that presents a paradigm shift from improvements have been small and commercial applica- perceptual (auditory) speech processing towards cognitive tions (e.g., cell phones) have been mostly satisfactory for (auditory plus cortical) speech processing. The objective the general public, and the growth of available bandwidth of this tutorial is to provide an overview of the impact has reduced requirements to compress speech even fur- of cognitive speech processing on speech compression and discuss challenges faced in this interdisciplinary speech ther.
    [Show full text]
  • A Novel Speech Enhancement Approach Based on Modified Dct and Improved Pitch Synchronous Analysis
    American Journal of Applied Sciences 11 (1): 24-37, 2014 ISSN: 1546-9239 ©2014 Science Publication doi:10.3844/ajassp.2014.24.37 Published Online 11 (1) 2014 (http://www.thescipub.com/ajas.toc) A NOVEL SPEECH ENHANCEMENT APPROACH BASED ON MODIFIED DCT AND IMPROVED PITCH SYNCHRONOUS ANALYSIS 1Balaji, V.R. and 2S. Subramanian 1Department of ECE, Sri Krishna College of Engineering and Technology, Coimbatore, India 2Department of CSE, Coimbatore Institute of Engineering and Technology, Coimbatore, India Received 2013-06-03, Revised 2013-07-17; Accepted 2013-11-21 ABSTRACT Speech enhancement has become an essential issue within the field of speech and signal processing, because of the necessity to enhance the performance of voice communication systems in noisy environment. There has been a number of research works being carried out in speech processing but still there is always room for improvement. The main aim is to enhance the apparent quality of the speech and to improve the intelligibility. Signal representation and enhancement in cosine transformation is observed to provide significant results. Discrete Cosine Transformation has been widely used for speech enhancement. In this research work, instead of DCT, Advanced DCT (ADCT) which simultaneous offers energy compaction along with critical sampling and flexible window switching. In order to deal with the issue of frame to frame deviations of the Cosine Transformations, ADCT is integrated with Pitch Synchronous Analysis (PSA). Moreover, in order to improve the noise minimization performance of the system, Improved Iterative Wiener Filtering approach called Constrained Iterative Wiener Filtering (CIWF) is used in this approach. Thus, a novel ADCT based speech enhancement using improved iterative filtering algorithm integrated with PSA is used in this approach.
    [Show full text]
  • Speech Compression
    information Review Speech Compression Jerry D. Gibson Department of Electrical and Computer Engineering, University of California, Santa Barbara, CA 93118, USA; [email protected]; Tel.: +1-805-893-6187 Academic Editor: Khalid Sayood Received: 22 April 2016; Accepted: 30 May 2016; Published: 3 June 2016 Abstract: Speech compression is a key technology underlying digital cellular communications, VoIP, voicemail, and voice response systems. We trace the evolution of speech coding based on the linear prediction model, highlight the key milestones in speech coding, and outline the structures of the most important speech coding standards. Current challenges, future research directions, fundamental limits on performance, and the critical open problem of speech coding for emergency first responders are all discussed. Keywords: speech coding; voice coding; speech coding standards; speech coding performance; linear prediction of speech 1. Introduction Speech coding is a critical technology for digital cellular communications, voice over Internet protocol (VoIP), voice response applications, and videoconferencing systems. In this paper, we present an abridged history of speech compression, a development of the dominant speech compression techniques, and a discussion of selected speech coding standards and their performance. We also discuss the future evolution of speech compression and speech compression research. We specifically develop the connection between rate distortion theory and speech compression, including rate distortion bounds for speech codecs. We use the terms speech compression, speech coding, and voice coding interchangeably in this paper. The voice signal contains not only what is said but also the vocal and aural characteristics of the speaker. As a consequence, it is usually desired to reproduce the voice signal, since we are interested in not only knowing what was said, but also in being able to identify the speaker.
    [Show full text]
  • Meeting Abstracts
    PROGRAM OF The 118th Meeting of the AcousticalSociety of America Adam's Mark Hotel ß St. Louis, Missouri ß 27 November-1 December 1989 MONDAY EVENING, 27 NOVEMBER 1989 ST. LOUIS BALLROOM D, 7:00 TO 9:00 P.M. Tutorial on Architectural Acoustics Mauro Pierucci, Chairman Departmentof Aerospaceand EngineeringMechanics, San DiegoState University,San Diego, California 92182 TUI. Architecturalacoustics: The forgottendimension. Ewart A. Wetherill (Wilson, lhrig, and Associates, lnc., 5776 Broadway,Oakland, CA 94618) The basic considerationsof architeclural acouslics---isolation from unwanled noise and vibration, control of mechanicalsystem noise, and room acousticsdesign---are all clearly exemplifiedin Sabinc'sdesign for BostonSymphony Hall. Openedin ! 900,this hall isone of theoutstanding successes in musical acoustics. Yet, aswe approachthe hundredthanniversary of Sabine'sfirst experiments, acoustical characteristics remain one of the leastconsidered aspects of buildingdesign. This is due, in part, to the difficultyof visualizingthe acouslicaloutcome of designdecisions, complicated by individualjudgment as to whatconstitutes good acous- tics. However,the lack of a comprehensiveteaching program remains the dominantproblem. Significant advancesover the past 2 or 3 decadesin measurementand evaluationhave refinedthe ability to design predictabilityand to demonsIrateacoustical concerns to others. New techniquessuch as sound intensity measurements,new descriptors for roomacoustics phenomena, and the refinemen t of recording,analysis, and amplificationtechniques provide fresh insights into the behaviorof soundin air and other media.These topics are reviewedwith particularemphasis on the needfor a comparableadvance in translationof acousticprinci- plesinto buildingtechnologies. Sl J. Acoust.Soc. Am. Suppl. 1, VoL86, Fall1989 118thMeeting: Acoustical Society of America S1 TUESDAY MORNING, 28 NOVEMBER 1989 ST. LOUIS BALLROOM C, 8:00 A.M. TO 12:00 NOON SessionA.
    [Show full text]
  • 4: Speech Compression
    4: Speech Compression Mark Handley Data Rates Telephone quality voice: 8000 samples/sec, 8 bits/sample, mono 64Kb/s CD quality audio: 44100 samples/sec, 16 bits/sample, stereo ~1.4Mb/s Communications channels and storage cost money (although less than they used to) What can we do to reduce the transmission and/or storage costs without sacrificing too much quality? 1 Speech Codec Overview PCM - send every sample DPCM - send differences between samples ADPCM - send differences, but adapt how we code them SB-ADPCM - wideband codec, use ADPCM twice, once for lower frequencies, again at lower bitrate for upper frequencies. LPC - linear model of speech formation CELP - use LPC as base, but also use some bits to code corrections for the things LPC gets wrong. PCM µ-law and a-law PCM have already reduced the data sent. Lost frequencies above 4KHz. Non-linear encoding to reduce bits per sample. However, each sample is still independently encoded. In reality, samples are correlated. Can utilize this correlation to reduce the data sent. 2 Differential PCM Normally the difference between samples is relatively small and can be coded with less than 8 bits. Simplest codec sends only the differences between samples. Typically use 6 bits for difference, rather than 8 bits for absolute value. Compression is lossy, as not all differences can be coded Decoded signal is slightly degraded. Next difference must then be encoded off the previous decoded sample, so losses don’t accumulate. Differential PCM Difference accurately coded Code next in 6 bits difference Lossy off previous Compression decoded sample max 6 bit difference 3 ADPCM (Adaptive Differential PCM) Makes a simple prediction of the next sample, based on weighted previous n samples.
    [Show full text]
  • Source Coding Basics and Speech Coding
    Source Coding Basics and Speech Coding Yao Wang Polytechnic University, Brooklyn, NY11201 http://eeweb.poly.edu/~yao Outline • Why do we need to compress speech signals • Basic components in a source coding system • Variable length binary encoding: Huffman coding • Speech coding overview • Predictive coding: DPCM, ADPCM • Vocoder • Hybrid coding • Speech coding standards ©Yao Wang, 2006 EE3414: Speech Coding 2 Why do we need to compress? • Raw PCM speech (sampled at 8 kbps, represented with 8 bit/sample) has data rate of 64 kbps • Speech coding refers to a process that reduces the bit rate of a speech file • Speech coding enables a telephone company to carry more voice calls in a single fiber or cable • Speech coding is necessary for cellular phones, which has limited data rate for each user (<=16 kbps is desired!). • For the same mobile user, the lower the bit rate for a voice call, the more other services (data/image/video) can be accommodated. • Speech coding is also necessary for voice-over-IP, audio-visual teleconferencing, etc, to reduce the bandwidth consumption over the Internet ©Yao Wang, 2006 EE3414: Speech Coding 3 Basic Components in a Source Coding System Input Transformed Quantized Binary Samples parameters parameters bitstreams Transfor- Quanti- Binary mation zation Encoding Lossy Lossless Prediction Scalar Q Fixed length Transforms Vector Q Variable length Model fitting (Huffman, …... arithmetic, LZW) • Motivation for transformation --- To yield a more efficient representation of the original samples. ©Yao Wang, 2006 EE3414:
    [Show full text]
  • Answers to Exercises
    Answers to Exercises A bird does not sing because he has an answer, he sings because he has a song. —Chinese Proverb Intro.1: abstemious, abstentious, adventitious, annelidous, arsenious, arterious, face- tious, sacrilegious. Intro.2: When a software house has a popular product they tend to come up with new versions. A user can update an old version to a new one, and the update usually comes as a compressed file on a floppy disk. Over time the updates get bigger and, at a certain point, an update may not fit on a single floppy. This is why good compression is important in the case of software updates. The time it takes to compress and decompress the update is unimportant since these operations are typically done just once. Recently, software makers have taken to providing updates over the Internet, but even in such cases it is important to have small files because of the download times involved. 1.1: (1) ask a question, (2) absolutely necessary, (3) advance warning, (4) boiling hot, (5) climb up, (6) close scrutiny, (7) exactly the same, (8) free gift, (9) hot water heater, (10) my personal opinion, (11) newborn baby, (12) postponed until later, (13) unexpected surprise, (14) unsolved mysteries. 1.2: A reasonable way to use them is to code the five most-common strings in the text. Because irreversible text compression is a special-purpose method, the user may know what strings are common in any particular text to be compressed. The user may specify five such strings to the encoder, and they should also be written at the start of the output stream, for the decoder’s use.
    [Show full text]
  • Influence of Statistical Surface Models on Dynamic Scattering of High-Frequency Signals from the Ocean Surface (A)
    Downloaded from orbit.dtu.dk on: Dec 17, 2017 Influence of statistical surface models on dynamic scattering of high-frequency signals from the ocean surface (A) Bjerrum-Niese, Christian; Jensen, Leif Bjørnø Published in: Acoustical Society of America. Journal Link to article, DOI: 10.1121/1.411137 Publication date: 1994 Document Version Publisher's PDF, also known as Version of record Link back to DTU Orbit Citation (APA): Bjerrum-Niese, C., & Jensen, L. B. (1994). Influence of statistical surface models on dynamic scattering of high- frequency signals from the ocean surface (A). Acoustical Society of America. Journal, 96(5), 3232-3232. DOI: 10.1121/1.411137 General rights Copyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights. • Users may download and print one copy of any publication from the public portal for the purpose of private study or research. • You may not further distribute the material or use it for any profit-making activity or commercial gain • You may freely distribute the URL identifying the publication in the public portal If you believe that this document breaches copyright please contact us providing details, and we will remove access to the work immediately and investigate your claim. PROGRAM OF The 128th Meeting of the AcousticalSociety of America Stouffer Austin Hotel ß Austin, Texas ß 28 November-2 December 1994 NOTE: All Journalarticles and Lettersto the Editor are peer reviewedbefore publication.
    [Show full text]