Digital Speech Processing— Lecture 17

Total Page:16

File Type:pdf, Size:1020Kb

Digital Speech Processing— Lecture 17 Digital Speech Processing— Lecture 17 Speech Coding Methods Based on Speech Models 1 Waveform Coding versus Block Processing • Waveform coding – sample-by-sample matching of waveforms – coding quality measured using SNR • Source modeling (block processing) – block processing of signal => vector of outputs every block – overlapped blocks Block 1 Block 2 Block 3 2 Model-Based Speech Coding • we’ve carried waveform coding based on optimizing and maximizing SNR about as far as possible – achieved bit rate reductions on the order of 4:1 (i.e., from 128 Kbps PCM to 32 Kbps ADPCM) at the same time achieving toll quality SNR for telephone-bandwidth speech • to lower bit rate further without reducing speech quality, we need to exploit features of the speech production model, including: – source modeling – spectrum modeling – use of codebook methods for coding efficiency • we also need a new way of comparing performance of different waveform and model-based coding methods – an objective measure, like SNR, isn’t an appropriate measure for model- based coders since they operate on blocks of speech and don’t follow the waveform on a sample-by-sample basis – new subjective measures need to be used that measure user-perceived quality, intelligibility, and robustness to multiple factors 3 Topics Covered in this Lecture • Enhancements for ADPCM Coders – pitch prediction – noise shaping • Analysis-by-Synthesis Speech Coders – multipulse linear prediction coder (MPLPC) – code-excited linear prediction (CELP) • Open-Loop Speech Coders – two-state excitation model – LPC vocoder – residual-excited linear predictive coder – mixed excitation systems • speech coding quality measures - MOS • speech coding standards 4 Differential Quantization x[n] d[n] dˆ[n] c[n] ~x[n] xˆ[n] P: simple predictor of vocal tract response c′[n ] dˆ ′[n ] xˆ ′[n ] p − k ~ P ( z ) = ∑ z x ′[n ] 5 k =1 Issues with Differential Quantization • difference signal retains the character of the excitation signal – switches back and forth between quasi- periodic and noise-like signals • prediction duration (even when using p=20) is order of 2.5 msec (for sampling rate of 8 kHz) – predictor is predicting vocal tract response – not the excitation period (for voiced sounds) • Solution – incorporate two stages of prediction, namely a short-time predictor for the vocal tract response and a long- time predictor for pitch period 6 Pitch Prediction ˆ ˆ xn[] dn[] Q[ ] dn[] xn[] + + + + + + − + + + + xn%[] + P1(z) P2(z) xnˆ[] P (z) P (z) = β z−M + + 1 1 + p Pc (z) −k P2(z) = ∑αk z P2(z) 1− P1(z) k=1 residual Transmitter Receiver • first stage pitch predictor: −M Pz1()= β ⋅ z • second stage linear predictor (vocal tract predictor): p −k Pz2 ()= ∑αk z 7 k =1 Pitch Prediction first stage pitch predictor: −M Pz1()=⋅β z this predictor model assumes that the pitch period, M, is an integer number of samples and β is a gain constant allowing for variations in pitch period over time (for unvoiced or background frames, values of M and β are irrelevant) an alternative (somewhat more complicated) pitch predictor is of the form: 1 −+MM1 − −MMk−−−1 Pz11()=++βββ z 2 z 3 z = ∑ βk z k =−1 this more advanced form provides a way to handle non-integer pitch period through interpolation around the nearest integer pitch period value, M 8 Combined Prediction The combined inverse system is the cascade in the decoder system: ⎛⎞⎛⎞11⎛⎞ 1 Hzc ()==⎜⎟⎜⎟⎜⎟ ⎝⎠⎝⎠1()1()1()−−Pz12 Pz⎝⎠ − Pzc with 2-stage prediction error filter of the form: 1−=−−Pzc ()1()1()11[][] Pz12 Pz =−−[]Pz12() Pz ()− Pz 1 () Pzc ()=−[] 1 Pz12 () Pz () − Pz 1 () which is implemented as a parallel combination of two predictors: []1()()− Pz12 Pz and Pz 1 () The prediction signal, xn%[] can be expressed as p ˆ ˆˆ xn%[]= β x[]nM−+∑αβk ([][ xnk −− xnkM −− ]) k =1 9 Combined Prediction Error The combined prediction error can be defined as: dnc[]=− xn [] xn% [] p =−vn[]∑αk vn [ − k ] k =1 where vn[]=− xn []β xn [ − M ] is the prediction error of the pitch predictor. The optimal values of β,M and {αk } are obtained, in theory, by minimizing the variance of dnc[]. In practice a sub-optimum solution is obtained by first minimizing the variance of vn[] and then minimizing the variance of dnc[] subject to the chosen values of β and M 10 Solution for Combined Predictor Mean-squared prediction error for pitch predictor is: 2 2 EvnxnxnM1 ==−−([])() []β [ ] where denotes averaging over a finite frame of speech samples. We use the covariance-type of averaging to eliminate windowing effects, giving the solution: xnxn[][− M ] β = ()opt 2 ()xn[]− M Using this value of β, we solve for E as ()1 opt ⎛⎞ ([][xnxn− M ])2 Exn=−[]2 ⎜⎟ 1 ()1 opt ( ) 22 ⎜⎟xn[] xn [− M ] ⎝⎠()() with minimum normalized covariance: xnxn[][− M ] ρ[]M = 1/2 xn[]22 xn [− M ] ( ()()) 11 Solution for Combined Predictor • Steps in solution: – first search for M that maximizes ρ[M] – compute βopt • Solve for more accurate pitch predictor by minimizing the variance of the expanded pitch predictor • Solve for optimum vocal tract predictor coefficients, αk, k=1,2,…,p 12 Pitch Prediction Vocal tract prediction Pitch and vocal tract prediction 13 Noise Shaping in DPCM Systems 14 Noise Shaping Fundamentals The output of an ADPCN encoder/decoder is: xnˆ[]=+ xn [] en [] where en[] is the quantization noise. It is easy to show that en[] generally has a flat spectrum and thus is especially audible in spectral regions of low intensity, i.e., between formants. This has led to methods of shaping the quantization noise to match the speech spectrum and take advantage of spectral masking concepts 15 Noise Shaping Basic ADPCM encoder and decoder Equivalent ADPCM encoder and decoder Noise shaping ADPCM encoder and decoder 16 Noise Shaping The equivalence of parts (b) and (a) is shown by the following: xnˆ[]=+↔ xn [] en [] Xzˆ () = Xz () + Ez () From part (a) we see that: Dz()=− Xz () PzXz ()ˆ () =− [1 PzXz ()]() − PzEz () () with Ez()=− Dzˆ () Dz () showing the equivalence of parts (b) and (a). Further, since Dzˆ ()=+=− Dz () Ez () [1 Pz ()]() X z +− [1 Pz ()]() Ez ˆˆ⎛⎞1 ˆ X() z== HzDz () ()⎜⎟ Dz () ⎝⎠1()− Pz ⎛⎞1 =−+−⎜⎟([1Pz ( )] X ( z ) [1 Pz ( )] Ez ( )) ⎝⎠1()− Pz =+Xz() Ez () Feeding back the quantization error through the predictor, Pz() ensures that the reconstructed signal, xnˆ [], differs from xn[], by the quantization error, en [], incurred in quantizing the difference signal, dn[]. 17 Shaping the Quantization Noise To shape the quantization noise we simply replace Pz() by a different system function, Fz(), giving the reconstructed signal as: ˆˆ⎛⎞1 ˆ X'( z )== HzDz ( ) '( )⎜⎟ Dz '( ) ⎝⎠1()− Pz ⎛⎞1 =−+−⎜⎟()[]1()()1(Pz X z[] FzEz)'() ⎝⎠1()− Pz ⎛⎞1()− Fz =+Xz()⎜⎟ E '() z ⎝⎠1()− Pz Thus if xn[] is coded by the encoder, the z-transform of the reconstructed signal at the receiver is: Xzˆ '( )=+ Xz ( ) Ez% '( ) ⎛⎞1()− Fz Ez% '()==Γ⎜⎟ Ez '() () zEz '() ⎝⎠1()− Pz ⎛⎞1()− Fz where Γ=()z ⎜⎟ is the effective noise shaping filter 18 ⎝⎠1()− Pz Noise Shaping Filter Options Noise shaping filter options: 1. Fz()= 0 and we assume noise has a flat spectrum, then the noise and speech spectrum have the same shape 2. Fz()= Pz () then the equivalent system is the standard DPCM system where Ez% '( )== Ez '( ) Ez ( ) with flat noise spectrum p −−1 kk 3.Fz ( )== P (γαγ z ) ∑ k z and we ''shape'' the noise spectrum k =1 to ''hide'' the noise beneath the spectral peaks of the speech signal; each zero of [1−−Pz ( )] is paired with a zero of [1 Fz ( )] where the paired zeros have the same angles in the z-plane, but with a radius that is divided by γ. 19 Noise Shaping Filter 20 Noise Shaping Filter If we assume that the quantization noise has a flat spectrum with noise 2 power of σ e' , then the power spectrum of the shaped noise is of the form: jFF2/π S jFF2/π S 1(− Fe )2 Peee''()= σ 1(− PejFF2/π S ) Noise spectrum above speech spectrum 21 Fully Quantized Adaptive Predictive Coder 22 Full ADPCM Coder • Input is x[n] • P2(z) is the short-term (vocal tract) predictor • Signal v[n] is the short-term prediction error • Goal of encoder is to obtain a quantized representation of this excitation signal, from which the original signal can be reconstructed. 23 Quantized ADPCM Coder Total bit rate for ADPCM coder: IBFBFBFADPCM=+ SΔΔ + P P where B is the number of bits for the quantization of the difference signal, BΔ is the number of bits for encoding the step size at frame rate FBΔ , and P is the total number of bits allocated to the predictor coefficients (both long and short-term) with frame rate FP Typically FBS =≈−8000 and even with 1 4 bits, we need between 8000 and 3200 bps for quantization of difference signal Typically we need about 3000-4000 bps for the side information (step size and predictor coefficients) Overall we need between 11,000 and 36,000 bps for a fully quantized system 24 Bit Rate for LP Coding • speech and residual sampling rate: Fs=8 kHz • LP analysis frame rate: F∆=FP = 50-100 frames/sec • quantizer stepsize: 6 bits/frame • predictor parameters: – M (pitch period): 7 bits/frame – pitch predictor coefficients: 13 bits/frame – vocal tract predictor coefficients: PARCORs 16-20, 46- 50 bits/frame • prediction residual: 1-3 bits/sample • total bit rate: – BR = 72*FP + Fs (minimum) 25 Two-Level (B=1 bit) Quantizer Prediction residual Quantizer input Quantizer output Reconstructed pitch Original pitch residual Reconstructed speech Original speech 26 Three-Level Center-Clipped Quantizer Prediction residual Quantizer input Quantizer output Reconstructed pitch Original pitch residual Reconstructed speech
Recommended publications
  • Speech Coding and Compression
    Introduction to voice coding/compression PCM coding Speech vocoders based on speech production LPC based vocoders Speech over packet networks Speech coding and compression Corso di Networked Multimedia Systems Master Universitario di Primo Livello in Progettazione e Gestione di Sistemi di Rete Carlo Drioli Università degli Studi di Verona Facoltà di Scienze Matematiche, Dipartimento di Informatica Fisiche e Naturali Speech coding and compression Introduction to voice coding/compression PCM coding Speech vocoders based on speech production LPC based vocoders Speech over packet networks Speech coding and compression: OUTLINE Introduction to voice coding/compression PCM coding Speech vocoders based on speech production LPC based vocoders Speech over packet networks Speech coding and compression Introduction to voice coding/compression PCM coding Speech vocoders based on speech production LPC based vocoders Speech over packet networks Introduction Approaches to voice coding/compression I Waveform coders (PCM) I Voice coders (vocoders) Quality assessment I Intelligibility I Naturalness (involves speaker identity preservation, emotion) I Subjective assessment: Listening test, Mean Opinion Score (MOS), Diagnostic acceptability measure (DAM), Diagnostic Rhyme Test (DRT) I Objective assessment: Signal to Noise Ratio (SNR), spectral distance measures, acoustic cues comparison Speech coding and compression Introduction to voice coding/compression PCM coding Speech vocoders based on speech production LPC based vocoders Speech over packet networks
    [Show full text]
  • Audio Coding for Digital Broadcasting
    Recommendation ITU-R BS.1196-7 (01/2019) Audio coding for digital broadcasting BS Series Broadcasting service (sound) ii Rec. ITU-R BS.1196-7 Foreword The role of the Radiocommunication Sector is to ensure the rational, equitable, efficient and economical use of the radio- frequency spectrum by all radiocommunication services, including satellite services, and carry out studies without limit of frequency range on the basis of which Recommendations are adopted. The regulatory and policy functions of the Radiocommunication Sector are performed by World and Regional Radiocommunication Conferences and Radiocommunication Assemblies supported by Study Groups. Policy on Intellectual Property Right (IPR) ITU-R policy on IPR is described in the Common Patent Policy for ITU-T/ITU-R/ISO/IEC referenced in Resolution ITU-R 1. Forms to be used for the submission of patent statements and licensing declarations by patent holders are available from http://www.itu.int/ITU-R/go/patents/en where the Guidelines for Implementation of the Common Patent Policy for ITU-T/ITU-R/ISO/IEC and the ITU-R patent information database can also be found. Series of ITU-R Recommendations (Also available online at http://www.itu.int/publ/R-REC/en) Series Title BO Satellite delivery BR Recording for production, archival and play-out; film for television BS Broadcasting service (sound) BT Broadcasting service (television) F Fixed service M Mobile, radiodetermination, amateur and related satellite services P Radiowave propagation RA Radio astronomy RS Remote sensing systems S Fixed-satellite service SA Space applications and meteorology SF Frequency sharing and coordination between fixed-satellite and fixed service systems SM Spectrum management SNG Satellite news gathering TF Time signals and frequency standards emissions V Vocabulary and related subjects Note: This ITU-R Recommendation was approved in English under the procedure detailed in Resolution ITU-R 1.
    [Show full text]
  • Mpeg Vbr Slice Layer Model Using Linear Predictive Coding and Generalized Periodic Markov Chains
    MPEG VBR SLICE LAYER MODEL USING LINEAR PREDICTIVE CODING AND GENERALIZED PERIODIC MARKOV CHAINS Michael R. Izquierdo* and Douglas S. Reeves** *Network Hardware Division IBM Corporation Research Triangle Park, NC 27709 [email protected] **Electrical and Computer Engineering North Carolina State University Raleigh, North Carolina 27695 [email protected] ABSTRACT The ATM Network has gained much attention as an effective means to transfer voice, video and data information We present an MPEG slice layer model for VBR over computer networks. ATM provides an excellent vehicle encoded video using Linear Predictive Coding (LPC) and for video transport since it provides low latency with mini- Generalized Periodic Markov Chains. Each slice position mal delay jitter when compared to traditional packet net- within an MPEG frame is modeled using an LPC autoregres- works [11]. As a consequence, there has been much research sive function. The selection of the particular LPC function is in the area of the transmission and multiplexing of com- governed by a Generalized Periodic Markov Chain; one pressed video data streams over ATM. chain is defined for each I, P, and B frame type. The model is Compressed video differs greatly from classical packet sufficiently modular in that sequences which exclude B data sources in that it is inherently quite bursty. This is due to frames can eliminate the corresponding Markov Chain. We both temporal and spatial content variations, bounded by a show that the model matches the pseudo-periodic autocorre- fixed picture display rate. Rate control techniques, such as lation function quite well. We present simulation results of CBR (Constant Bit Rate), were developed in order to reduce an Asynchronous Transfer Mode (ATM) video transmitter the burstiness of a video stream.
    [Show full text]
  • Speech Compression Using Discrete Wavelet Transform and Discrete Cosine Transform
    International Journal of Engineering Research & Technology (IJERT) ISSN: 2278-0181 Vol. 1 Issue 5, July - 2012 Speech Compression Using Discrete Wavelet Transform and Discrete Cosine Transform Smita Vatsa, Dr. O. P. Sahu M. Tech (ECE Student ) Professor Department of Electronicsand Department of Electronics and Communication Engineering Communication Engineering NIT Kurukshetra, India NIT Kurukshetra, India Abstract reproduced with desired level of quality. Main approaches of speech compression used today are Aim of this paper is to explain and implement waveform coding, transform coding and parametric transform based speech compression techniques. coding. Waveform coding attempts to reproduce Transform coding is based on compressing signal input signal waveform at the output. In transform by removing redundancies present in it. Speech coding at the beginning of procedure signal is compression (coding) is a technique to transform transformed into frequency domain, afterwards speech signal into compact format such that speech only dominant spectral features of signal are signal can be transmitted and stored with reduced maintained. In parametric coding signals are bandwidth and storage space respectively represented through a small set of parameters that .Objective of speech compression is to enhance can describe it accurately. Parametric coders transmission and storage capacity. In this paper attempts to produce a signal that sounds like Discrete wavelet transform and Discrete cosine original speechwhether or not time waveform transform
    [Show full text]
  • Speech Coding Using Code Excited Linear Prediction
    ISSN : 0976-8491 (Online) | ISSN : 2229-4333 (Print) IJCST VOL . 5, Iss UE SPL - 2, JAN - MAR C H 2014 Speech Coding Using Code Excited Linear Prediction 1Hemanta Kumar Palo, 2Kailash Rout 1ITER, Siksha ‘O’ Anusandhan University, Bhubaneswar, Odisha, India 2Gandhi Institute For Technology, Bhubaneswar, Odisha, India Abstract The main problem with the speech coding system is the optimum utilization of channel bandwidth. Due to this the speech signal is coded by using as few bits as possible to get low bit-rate speech coders. As the bit rate of the coder goes low, the intelligibility, SNR and overall quality of the speech signal decreases. Hence a comparative analysis is done of two different types of speech coders in this paper for understanding the utility of these coders in various applications so as to reduce the bandwidth and by different speech coding techniques and by reducing the number of bits without any appreciable compromise on the quality of speech. Hindi language has different number of stops than English , hence the performance of the coders must be checked on different languages. The main objective of this paper is to develop speech coders capable of producing high quality speech at low data rates. The focus of this paper is the development and testing of voice coding systems which cater for the above needs. Keywords PCM, DPCM, ADPCM, LPC, CELP Fig. 1: Speech Production Process I. Introduction Speech coding or speech compression is the compact digital The two types of speech sounds are voiced and unvoiced [1]. representations of voice signals [1-3] for the purpose of efficient They produce different sounds and spectra due to their differences storage and transmission.
    [Show full text]
  • An Ultra Low-Power Miniature Speech CODEC at 8 Kb/S and 16 Kb/S Robert Brennan, David Coode, Dustin Griesdorf, Todd Schneider Dspfactory Ltd
    To appear in ICSPAT 2000 Proceedings, October 16-19, 2000, Dallas, TX. An Ultra Low-power Miniature Speech CODEC at 8 kb/s and 16 kb/s Robert Brennan, David Coode, Dustin Griesdorf, Todd Schneider Dspfactory Ltd. 611 Kumpf Drive, Unit 200 Waterloo, Ontario, N2V 1K8 Canada Abstract SmartCODEC platform consists of an effi- cient, block-floating point, oversampled This paper describes a CODEC implementa- Weighted OverLap-Add (WOLA) filterbank, a tion on an ultra low-power miniature applica- software-programmable dual-Harvard 16-bit tion specific signal processor (ASSP) designed DSP core, two high fidelity 14-bit A/D con- for mobile audio signal processing applica- verters, a 14-bit D/A converter and a flexible tions. The CODEC records speech to and set of peripherals [1]. The system hardware plays back from a serial flash memory at data architecture (Figure 1) was designed to enable rates of 16 and 8 kb/s, with a bandwidth of 4 memory upgrades. Removable memory cards kHz. This CODEC consumes only 1 mW in a or more power-efficient memory could be package small enough for use in a range of substituted for the serial flash memory. demanding portable applications. Results, improvements and applications are also dis- The CODEC communicates with the flash cussed. memory over an integrated SPI port that can transfer data at rates up to 80 kb/s. For this application, the port is configured to block 1. Introduction transfer frame packets every 14 ms. e Speech coding is ubiquitous. Increasing de- c i WOLA Filterbank v mands for portability and fidelity coupled with A/D e and D the desire for reduced storage and bandwidth o i l Programmable d o r u utilization have increased the demand for and t D/A DSP Core A deployment of speech CODECs.
    [Show full text]
  • TMS320C6000 U-Law and A-Law Companding with Software Or The
    Application Report SPRA634 - April 2000 TMS320C6000t µ-Law and A-Law Companding with Software or the McBSP Mark A. Castellano Digital Signal Processing Solutions Todd Hiers Rebecca Ma ABSTRACT This document describes how to perform data companding with the TMS320C6000 digital signal processors (DSP). Companding refers to the compression and expansion of transfer data before and after transmission, respectively. The multichannel buffered serial port (McBSP) in the TMS320C6000 supports two companding formats: µ-Law and A-Law. Both companding formats are specified in the CCITT G.711 recommendation [1]. This application report discusses how to use the McBSP to perform data companding in both the µ-Law and A-Law formats. This document also discusses how to perform companding of data not coming into the device but instead located in a buffer in processor memory. Sample McBSP setup codes are provided for reference. In addition, this application report discusses µ-Law and A-Law companding in software. The appendix provides software companding codes and a brief discussion on the companding theory. Contents 1 Design Problem . 2 2 Overview . 2 3 Companding With the McBSP. 3 3.1 µ-Law Companding . 3 3.1.1 McBSP Register Configuration for µ-Law Companding. 4 3.2 A-Law Companding. 4 3.2.1 McBSP Register Configuration for A-Law Companding. 4 3.3 Companding Internal Data. 5 3.3.1 Non-DLB Mode. 5 3.3.2 DLB Mode . 6 3.4 Sample C Functions. 6 4 Companding With Software. 6 5 Conclusion . 7 6 References . 7 TMS320C6000 is a trademark of Texas Instruments Incorporated.
    [Show full text]
  • Bit Rate Adaptation Using Linear Quadratic Optimization for Mobile Video Streaming
    applied sciences Article Bit Rate Adaptation Using Linear Quadratic Optimization for Mobile Video Streaming Young-myoung Kang 1 and Yeon-sup Lim 2,* 1 Network Business, Samsung Electronics, Suwon 16677, Korea; [email protected] 2 Department of Convergence Security Engineering, Sungshin Women’s University, Seoul 02844, Korea * Correspondence: [email protected]; Tel.: +82-2-920-7144 Abstract: Video streaming application such as Youtube is one of the most popular mobile applications. To adjust the quality of video for available network bandwidth, a streaming server provides multiple representations of video of which bit rate has different bandwidth requirements. A streaming client utilizes an adaptive bit rate (ABR) scheme to select a proper representation that the network can support. However, in mobile environments, incorrect decisions of an ABR scheme often cause playback stalls that significantly degrade the quality of user experience, which can easily happen due to network dynamics. In this work, we propose a control theory (Linear Quadratic Optimization)- based ABR scheme to enhance the quality of experience in mobile video streaming. Our simulation study shows that our proposed ABR scheme successfully mitigates and shortens playback stalls while preserving the similar quality of streaming video compared to the state-of-the-art ABR schemes. Keywords: dynamic adaptive streaming over HTTP; adaptive bit rate scheme; control theory 1. Introduction Video streaming constitutes a large fraction of current Internet traffic. In particular, video streaming accounts for more than 65% of world-wide mobile downstream traffic [1]. Commercial streaming services such as YouTube and Netflix are implemented based on Citation: Kang, Y.-m.; Lim, Y.-s.
    [Show full text]
  • Cognitive Speech Coding Milos Cernak, Senior Member, IEEE, Afsaneh Asaei, Senior Member, IEEE, Alexandre Hyafil
    1 Cognitive Speech Coding Milos Cernak, Senior Member, IEEE, Afsaneh Asaei, Senior Member, IEEE, Alexandre Hyafil Abstract—Speech coding is a field where compression ear and undergoes a highly complex transformation paradigms have not changed in the last 30 years. The before it is encoded efficiently by spikes at the auditory speech signals are most commonly encoded with com- nerve. This great efficiency in information representation pression methods that have roots in Linear Predictive has inspired speech engineers to incorporate aspects of theory dating back to the early 1940s. This paper tries to cognitive processing in when developing efficient speech bridge this influential theory with recent cognitive studies applicable in speech communication engineering. technologies. This tutorial article reviews the mechanisms of speech Speech coding is a field where research has slowed perception that lead to perceptual speech coding. Then considerably in recent years. This has occurred not it focuses on human speech communication and machine because it has achieved the ultimate in minimizing bit learning, and application of cognitive speech processing in rate for transparent speech quality, but because recent speech compression that presents a paradigm shift from improvements have been small and commercial applica- perceptual (auditory) speech processing towards cognitive tions (e.g., cell phones) have been mostly satisfactory for (auditory plus cortical) speech processing. The objective the general public, and the growth of available bandwidth of this tutorial is to provide an overview of the impact has reduced requirements to compress speech even fur- of cognitive speech processing on speech compression and discuss challenges faced in this interdisciplinary speech ther.
    [Show full text]
  • A Predictive H.263 Bit-Rate Control Scheme Based on Scene
    A PRJ3DICTIVE H.263 BIT-RATE CONTROL SCHEME BASED ON SCENE INFORMATION Pohsiang Hsu and K. J. Ray Liu Department of Electrical and Computer Engineering University of Maryland at College Park College Park, Maryland, USA ABSTRACT is typically constant bit rate (e.g. PSTN). Under this cir- cumstance, the encoder’s output bit-rate must be regulated For many real-time visual communication applications such to meet the transmission bandwidth in order to guaran- as video telephony, the transmission channel in used is typ- tee a constant frame rate and delay. Given the channel ically constant bit rate (e.g. PSTN). Under this circum- rate and a target frame rate, we derive the target bits per stance, the encoder’s output bit-rate must be regulated to frame by distribute the channel rate equally to each hme. meet the transmission bandwidth in order to guarantee a Ffate control is typically based on varying the quantization constant frame rate and delay. In this work, we propose a paraneter to meet the target bit budget for each frame. new predictive rate control scheme based on the activity of Many rate control scheme for the block-based DCT/motion the scene to be coded for H.263 encoder compensation video encoders have been proposed in the past. These rate control algorithms are typically focused on MPEG video coding standards. Rate-Distortion based rate 1. INTRODUCTION control for MPEG using Lagrangian optimization [4] have been proposed in the literature. These methods require ex- In the past few years, emerging demands for digital video tensive,rate-distortion measurements and are unsuitable for communication applications such as video conferencing, video real time applications.
    [Show full text]
  • A Novel Speech Enhancement Approach Based on Modified Dct and Improved Pitch Synchronous Analysis
    American Journal of Applied Sciences 11 (1): 24-37, 2014 ISSN: 1546-9239 ©2014 Science Publication doi:10.3844/ajassp.2014.24.37 Published Online 11 (1) 2014 (http://www.thescipub.com/ajas.toc) A NOVEL SPEECH ENHANCEMENT APPROACH BASED ON MODIFIED DCT AND IMPROVED PITCH SYNCHRONOUS ANALYSIS 1Balaji, V.R. and 2S. Subramanian 1Department of ECE, Sri Krishna College of Engineering and Technology, Coimbatore, India 2Department of CSE, Coimbatore Institute of Engineering and Technology, Coimbatore, India Received 2013-06-03, Revised 2013-07-17; Accepted 2013-11-21 ABSTRACT Speech enhancement has become an essential issue within the field of speech and signal processing, because of the necessity to enhance the performance of voice communication systems in noisy environment. There has been a number of research works being carried out in speech processing but still there is always room for improvement. The main aim is to enhance the apparent quality of the speech and to improve the intelligibility. Signal representation and enhancement in cosine transformation is observed to provide significant results. Discrete Cosine Transformation has been widely used for speech enhancement. In this research work, instead of DCT, Advanced DCT (ADCT) which simultaneous offers energy compaction along with critical sampling and flexible window switching. In order to deal with the issue of frame to frame deviations of the Cosine Transformations, ADCT is integrated with Pitch Synchronous Analysis (PSA). Moreover, in order to improve the noise minimization performance of the system, Improved Iterative Wiener Filtering approach called Constrained Iterative Wiener Filtering (CIWF) is used in this approach. Thus, a novel ADCT based speech enhancement using improved iterative filtering algorithm integrated with PSA is used in this approach.
    [Show full text]
  • Speech Compression
    information Review Speech Compression Jerry D. Gibson Department of Electrical and Computer Engineering, University of California, Santa Barbara, CA 93118, USA; [email protected]; Tel.: +1-805-893-6187 Academic Editor: Khalid Sayood Received: 22 April 2016; Accepted: 30 May 2016; Published: 3 June 2016 Abstract: Speech compression is a key technology underlying digital cellular communications, VoIP, voicemail, and voice response systems. We trace the evolution of speech coding based on the linear prediction model, highlight the key milestones in speech coding, and outline the structures of the most important speech coding standards. Current challenges, future research directions, fundamental limits on performance, and the critical open problem of speech coding for emergency first responders are all discussed. Keywords: speech coding; voice coding; speech coding standards; speech coding performance; linear prediction of speech 1. Introduction Speech coding is a critical technology for digital cellular communications, voice over Internet protocol (VoIP), voice response applications, and videoconferencing systems. In this paper, we present an abridged history of speech compression, a development of the dominant speech compression techniques, and a discussion of selected speech coding standards and their performance. We also discuss the future evolution of speech compression and speech compression research. We specifically develop the connection between rate distortion theory and speech compression, including rate distortion bounds for speech codecs. We use the terms speech compression, speech coding, and voice coding interchangeably in this paper. The voice signal contains not only what is said but also the vocal and aural characteristics of the speaker. As a consequence, it is usually desired to reproduce the voice signal, since we are interested in not only knowing what was said, but also in being able to identify the speaker.
    [Show full text]