MP3 and AAC Explained
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Digital Speech Processing— Lecture 17
Digital Speech Processing— Lecture 17 Speech Coding Methods Based on Speech Models 1 Waveform Coding versus Block Processing • Waveform coding – sample-by-sample matching of waveforms – coding quality measured using SNR • Source modeling (block processing) – block processing of signal => vector of outputs every block – overlapped blocks Block 1 Block 2 Block 3 2 Model-Based Speech Coding • we’ve carried waveform coding based on optimizing and maximizing SNR about as far as possible – achieved bit rate reductions on the order of 4:1 (i.e., from 128 Kbps PCM to 32 Kbps ADPCM) at the same time achieving toll quality SNR for telephone-bandwidth speech • to lower bit rate further without reducing speech quality, we need to exploit features of the speech production model, including: – source modeling – spectrum modeling – use of codebook methods for coding efficiency • we also need a new way of comparing performance of different waveform and model-based coding methods – an objective measure, like SNR, isn’t an appropriate measure for model- based coders since they operate on blocks of speech and don’t follow the waveform on a sample-by-sample basis – new subjective measures need to be used that measure user-perceived quality, intelligibility, and robustness to multiple factors 3 Topics Covered in this Lecture • Enhancements for ADPCM Coders – pitch prediction – noise shaping • Analysis-by-Synthesis Speech Coders – multipulse linear prediction coder (MPLPC) – code-excited linear prediction (CELP) • Open-Loop Speech Coders – two-state excitation -
Biographies of Computer Scientists
1 Charles Babbage 26 December 1791 (London, UK) – 18 October 1871 (London, UK) Life and Times Charles Babbage was born into a wealthy family, and started his mathematics education very early. By . 1811, when he went to Trinity College, Cambridge, he found that he knew more mathematics then his professors. He moved to Peterhouse, Cambridge from where he graduated in 1814. However, rather than come second to his friend Herschel in the final examinations, Babbage decided not to compete for an honors degree. In 1815 he co-founded the Analytical Society dedicated to studying continental reforms of Newton's formulation of “The Calculus”. He was one of the founders of the Astronomical Society in 1820. In 1821 Babbage started work on his Difference Engine designed to accurately compile tables. Babbage received government funding to construct an actual machine, but they stopped the funding in 1832 when it became clear that its construction was running well over-budget George Schuetz completed a machine based on the design of the Difference Engine in 1854. On completing the design of the Difference Engine, Babbage started work on the Analytical Engine capable of more general symbolic manipulations. The design of the Analytical Engine was complete in 1856, but a complete machine would not be constructed for over a century. Babbage's interests were wide. It is claimed that he invented cow-catchers for railway engines, the uniform postal rate, a means of recognizing lighthouses. He was also interested in locks and ciphers. He was politically active and wrote many treatises. One of the more famous proposed the banning of street musicians. -
Audio Coding for Digital Broadcasting
Recommendation ITU-R BS.1196-7 (01/2019) Audio coding for digital broadcasting BS Series Broadcasting service (sound) ii Rec. ITU-R BS.1196-7 Foreword The role of the Radiocommunication Sector is to ensure the rational, equitable, efficient and economical use of the radio- frequency spectrum by all radiocommunication services, including satellite services, and carry out studies without limit of frequency range on the basis of which Recommendations are adopted. The regulatory and policy functions of the Radiocommunication Sector are performed by World and Regional Radiocommunication Conferences and Radiocommunication Assemblies supported by Study Groups. Policy on Intellectual Property Right (IPR) ITU-R policy on IPR is described in the Common Patent Policy for ITU-T/ITU-R/ISO/IEC referenced in Resolution ITU-R 1. Forms to be used for the submission of patent statements and licensing declarations by patent holders are available from http://www.itu.int/ITU-R/go/patents/en where the Guidelines for Implementation of the Common Patent Policy for ITU-T/ITU-R/ISO/IEC and the ITU-R patent information database can also be found. Series of ITU-R Recommendations (Also available online at http://www.itu.int/publ/R-REC/en) Series Title BO Satellite delivery BR Recording for production, archival and play-out; film for television BS Broadcasting service (sound) BT Broadcasting service (television) F Fixed service M Mobile, radiodetermination, amateur and related satellite services P Radiowave propagation RA Radio astronomy RS Remote sensing systems S Fixed-satellite service SA Space applications and meteorology SF Frequency sharing and coordination between fixed-satellite and fixed service systems SM Spectrum management SNG Satellite news gathering TF Time signals and frequency standards emissions V Vocabulary and related subjects Note: This ITU-R Recommendation was approved in English under the procedure detailed in Resolution ITU-R 1. -
Convention Program
AAEESS 112244tthh CCOONNVVEENNTTIIOONN PPRROOGGRRAAMM May 17 – 20, 2008 RAI International Exhibition & Congress Centre, Amsterdam The AES has launched a new opportunity to recognize 1University of Salford, Salford, Greater student members who author technical papers. The Stu- Manchester, UK dent Paper Award Competition is based on the preprint 2ICW, Ltd., Wrexham, Wales, UK manuscripts accepted for the AES convention. Forty-two student-authored papers were nominated. This paper gives an account of work carried out The excellent quality of the submissions has made the to assess the effects of metallized film selection process both challenging and exhilarating. polypropylene crossover capacitors on key sonic The award-winning student paper will be honored dur- attributes of reproduced sound. The capacitors ing the convention, and the student-authored manuscript under investigation were found to be mechani- will be published in a timely manner in the Journal of the cally resonant within the audio frequency band, Audio Engineering Society. and results obtained from subjective listening Nominees for the Student Paper Award were required tests have shown this to have a measurable to meet the following qualifications: effect on audio delivery. The listening test methodology employed in this study evolved (a) The paper was accepted for presentation at the from initial ABX type tests with set program AES 124th Convention. material to the final A/B tests where trained test (b) The first author was a student when the work was subjects used program material that they were conducted and the manuscript prepared. familiar with. The main findings were that (c) The student author’s affiliation listed in the manu- capacitors used in crossover circuitry can exhibit script is an accredited educational institution. -
Preview - Click Here to Buy the Full Publication
This is a preview - click here to buy the full publication IEC 62481-2 ® Edition 2.0 2013-09 INTERNATIONAL STANDARD colour inside Digital living network alliance (DLNA) home networked device interoperability guidelines – Part 2: DLNA media formats INTERNATIONAL ELECTROTECHNICAL COMMISSION PRICE CODE XH ICS 35.100.05; 35.110; 33.160 ISBN 978-2-8322-0937-0 Warning! Make sure that you obtained this publication from an authorized distributor. ® Registered trademark of the International Electrotechnical Commission This is a preview - click here to buy the full publication – 2 – 62481-2 © IEC:2013(E) CONTENTS FOREWORD ......................................................................................................................... 20 INTRODUCTION ................................................................................................................... 22 1 Scope ............................................................................................................................. 23 2 Normative references ..................................................................................................... 23 3 Terms, definitions and abbreviated terms ....................................................................... 30 3.1 Terms and definitions ............................................................................................ 30 3.2 Abbreviated terms ................................................................................................. 34 3.4 Conventions ......................................................................................................... -
Subjective Assessment of Audio Quality – the Means and Methods Within the EBU
Subjective assessment of audio quality – the means and methods within the EBU W. Hoeg (Deutsch Telekom Berkom) L. Christensen (Danmarks Radio) R. Walker (BBC) This article presents a number of useful means and methods for the subjective quality assessment of audio programme material in radio and television, developed and verified by EBU Project Group, P/LIST. The methods defined in several new 1. Introduction EBU Recommendations and Technical Documents are suitable for The existing EBU Recommendation, R22 [1], both operational and training states that “the amount of sound programme ma- purposes in broadcasting terial which is exchanged between EBU Members, organizations. and between EBU Members and other production organizations, continues to increase” and that “the only sufficient method of assessing the bal- ance of features which contribute to the quality of An essential prerequisite for ensuring a uniform the sound in a programme is by listening to it.” high quality of sound programmes is to standard- ize the means and methods required for their as- Therefore, “listening” is an integral part of all sessment. The subjective assessment of sound sound and television programme-making opera- quality has for a long time been carried out by tions. Despite the very significant advances of international organizations such as the (former) modern sound monitoring and measurement OIRT [2][3][4], the (former) CCIR (now ITU-R) technology, these essentially objective solutions [5] and the Member organizations of the EBU remain unable to tell us what the programme will itself. It became increasingly important that com- really sound like to the listener at home. -
TMS320C6000 U-Law and A-Law Companding with Software Or The
Application Report SPRA634 - April 2000 TMS320C6000t µ-Law and A-Law Companding with Software or the McBSP Mark A. Castellano Digital Signal Processing Solutions Todd Hiers Rebecca Ma ABSTRACT This document describes how to perform data companding with the TMS320C6000 digital signal processors (DSP). Companding refers to the compression and expansion of transfer data before and after transmission, respectively. The multichannel buffered serial port (McBSP) in the TMS320C6000 supports two companding formats: µ-Law and A-Law. Both companding formats are specified in the CCITT G.711 recommendation [1]. This application report discusses how to use the McBSP to perform data companding in both the µ-Law and A-Law formats. This document also discusses how to perform companding of data not coming into the device but instead located in a buffer in processor memory. Sample McBSP setup codes are provided for reference. In addition, this application report discusses µ-Law and A-Law companding in software. The appendix provides software companding codes and a brief discussion on the companding theory. Contents 1 Design Problem . 2 2 Overview . 2 3 Companding With the McBSP. 3 3.1 µ-Law Companding . 3 3.1.1 McBSP Register Configuration for µ-Law Companding. 4 3.2 A-Law Companding. 4 3.2.1 McBSP Register Configuration for A-Law Companding. 4 3.3 Companding Internal Data. 5 3.3.1 Non-DLB Mode. 5 3.3.2 DLB Mode . 6 3.4 Sample C Functions. 6 4 Companding With Software. 6 5 Conclusion . 7 6 References . 7 TMS320C6000 is a trademark of Texas Instruments Incorporated. -
Lossless Compression of Audio Data
CHAPTER 12 Lossless Compression of Audio Data ROBERT C. MAHER OVERVIEW Lossless data compression of digital audio signals is useful when it is necessary to minimize the storage space or transmission bandwidth of audio data while still maintaining archival quality. Available techniques for lossless audio compression, or lossless audio packing, generally employ an adaptive waveform predictor with a variable-rate entropy coding of the residual, such as Huffman or Golomb-Rice coding. The amount of data compression can vary considerably from one audio waveform to another, but ratios of less than 3 are typical. Several freeware, shareware, and proprietary commercial lossless audio packing programs are available. 12.1 INTRODUCTION The Internet is increasingly being used as a means to deliver audio content to end-users for en tertainment, education, and commerce. It is clearly advantageous to minimize the time required to download an audio data file and the storage capacity required to hold it. Moreover, the expec tations of end-users with regard to signal quality, number of audio channels, meta-data such as song lyrics, and similar additional features provide incentives to compress the audio data. 12.1.1 Background In the past decade there have been significant breakthroughs in audio data compression using lossy perceptual coding [1]. These techniques lower the bit rate required to represent the signal by establishing perceptual error criteria, meaning that a model of human hearing perception is Copyright 2003. Elsevier Science (USA). 255 AU rights reserved. 256 PART III / APPLICATIONS used to guide the elimination of excess bits that can be either reconstructed (redundancy in the signal) orignored (inaudible components in the signal). -
Equalization
CHAPTER 3 Equalization 35 The most intuitive effect — OK EQ is EZ U need a Hi EQ IQ MIX SMART QUICK START: Equalization GOALS ■ Fix spectral problems such as rumble, hum and buzz, pops and wind, proximity, and hiss. ■ Fit things into the mix by leveraging of any spectral openings available and through complementary cuts and boosts on spectrally competitive tracks. ■ Feature those aspects of each instrument that players and music fans like most. GEAR ■ Master the user controls for parametric, semiparametric, program, graphic, shelving, high-pass and low-pass filters. ■ Choose the equalizer with the capabilities you need, focusing particularly on the parameters, slopes, and number of bands available. Home stereos have tone controls. We in the studio get equalizers. Perhaps bet- ter described as a “spectral modifier” or “frequency-specific amplitude adjuster,” the equalizer allows the mix engineer to increase or decrease the level of specific frequency ranges within a signal. Having an equalizer is like having multiple volume knobs for a single track. Unlike the volume knob that attenuates or boosts an entire signal, the equalizer is the tool used to turn down or turn up specific frequency portions of any audio track . Out of all signal-processing tools that we use while mixing and tracking, EQ is probably the easiest, most intuitive to use—at first. Advanced applications of equalization, however, demand a deep understanding of the effect and all its Mix Smart. © 2011 Elsevier Inc. All rights reserved. 36 Mix Smart possibilities. Don't underestimate the intellectual challenge and creative poten- tial of this essential mix processor. -
Bit Rate Adaptation Using Linear Quadratic Optimization for Mobile Video Streaming
applied sciences Article Bit Rate Adaptation Using Linear Quadratic Optimization for Mobile Video Streaming Young-myoung Kang 1 and Yeon-sup Lim 2,* 1 Network Business, Samsung Electronics, Suwon 16677, Korea; [email protected] 2 Department of Convergence Security Engineering, Sungshin Women’s University, Seoul 02844, Korea * Correspondence: [email protected]; Tel.: +82-2-920-7144 Abstract: Video streaming application such as Youtube is one of the most popular mobile applications. To adjust the quality of video for available network bandwidth, a streaming server provides multiple representations of video of which bit rate has different bandwidth requirements. A streaming client utilizes an adaptive bit rate (ABR) scheme to select a proper representation that the network can support. However, in mobile environments, incorrect decisions of an ABR scheme often cause playback stalls that significantly degrade the quality of user experience, which can easily happen due to network dynamics. In this work, we propose a control theory (Linear Quadratic Optimization)- based ABR scheme to enhance the quality of experience in mobile video streaming. Our simulation study shows that our proposed ABR scheme successfully mitigates and shortens playback stalls while preserving the similar quality of streaming video compared to the state-of-the-art ABR schemes. Keywords: dynamic adaptive streaming over HTTP; adaptive bit rate scheme; control theory 1. Introduction Video streaming constitutes a large fraction of current Internet traffic. In particular, video streaming accounts for more than 65% of world-wide mobile downstream traffic [1]. Commercial streaming services such as YouTube and Netflix are implemented based on Citation: Kang, Y.-m.; Lim, Y.-s. -
4. MPEG Layer-3 Audio Encoding
MPEG Layer-3 An introduction to MPEG Layer-3 K. Brandenburg and H. Popp Fraunhofer Institut für Integrierte Schaltungen (IIS) MPEG Layer-3, otherwise known as MP3, has generated a phenomenal interest among Internet users, or at least among those who want to download highly-compressed digital audio files at near-CD quality. This article provides an introduction to the work of the MPEG group which was, and still is, responsible for bringing this open (i.e. non-proprietary) compression standard to the forefront of Internet audio downloads. 1. Introduction The audio coding scheme MPEG Layer-3 will soon celebrate its 10th birthday, having been standardized in 1991. In its first years, the scheme was mainly used within DSP- based codecs for studio applications, allowing professionals to use ISDN phone lines as cost-effective music links with high sound quality. In 1995, MPEG Layer-3 was selected as the audio format for the digital satellite broadcasting system developed by World- Space. This was its first step into the mass market. Its second step soon followed, due to the use of the Internet for the electronic distribution of music. Here, the proliferation of audio material – coded with MPEG Layer-3 (aka MP3) – has shown an exponential growth since 1995. By early 1999, “.mp3” had become the most popular search term on the Web (according to http://www.searchterms.com). In 1998, the “MPMAN” (by Saehan Information Systems, South Korea) was the first portable MP3 player, pioneering the road for numerous other manufacturers of consumer electronics. “MP3” has been featured in many articles, mostly on the business pages of newspapers and periodicals, due to its enormous impact on the recording industry. -
The H.264 Advanced Video Coding (AVC) Standard
Whitepaper: The H.264 Advanced Video Coding (AVC) Standard What It Means to Web Camera Performance Introduction A new generation of webcams is hitting the market that makes video conferencing a more lifelike experience for users, thanks to adoption of the breakthrough H.264 standard. This white paper explains some of the key benefits of H.264 encoding and why cameras with this technology should be on the shopping list of every business. The Need for Compression Today, Internet connection rates average in the range of a few megabits per second. While VGA video requires 147 megabits per second (Mbps) of data, full high definition (HD) 1080p video requires almost one gigabit per second of data, as illustrated in Table 1. Table 1. Display Resolution Format Comparison Format Horizontal Pixels Vertical Lines Pixels Megabits per second (Mbps) QVGA 320 240 76,800 37 VGA 640 480 307,200 147 720p 1280 720 921,600 442 1080p 1920 1080 2,073,600 995 Video Compression Techniques Digital video streams, especially at high definition (HD) resolution, represent huge amounts of data. In order to achieve real-time HD resolution over typical Internet connection bandwidths, video compression is required. The amount of compression required to transmit 1080p video over a three megabits per second link is 332:1! Video compression techniques use mathematical algorithms to reduce the amount of data needed to transmit or store video. Lossless Compression Lossless compression changes how data is stored without resulting in any loss of information. Zip files are losslessly compressed so that when they are unzipped, the original files are recovered.