MPEG-1MPEG --11 Standardsstandards
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
A Multidimensional Companding Scheme for Source Coding Witha Perceptually Relevant Distortion Measure
A MULTIDIMENSIONAL COMPANDING SCHEME FOR SOURCE CODING WITHA PERCEPTUALLY RELEVANT DISTORTION MEASURE J. Crespo, P. Aguiar R. Heusdens Department of Electrical and Computer Engineering Department of Mediamatics Instituto Superior Técnico Technische Universiteit Delft Lisboa, Portugal Delft, The Netherlands ABSTRACT coding mechanisms based on quantization, where less percep- tually important information is quantized more roughly and In this paper, we develop a multidimensional companding vice-versa, and lossless coding of the symbols emitted by the scheme for the preceptual distortion measure by S. van de Par quantizer to reduce statistical redundancy. In the quantization [1]. The scheme is asymptotically optimal in the sense that process of the (en)coder, it is desirable to quantize the infor- it has a vanishing rate-loss with increasing vector dimension. mation extracted from the signal in a rate-distortion optimal The compressor operates in the frequency domain: in its sim- sense, i.e., in a way that minimizes the perceptual distortion plest form, it pointwise multiplies the Discrete Fourier Trans- experienced by the user subject to the constraint of a certain form (DFT) of the windowed input signal by the square-root available bitrate. of the inverse of the masking threshold, and then goes back into the time domain with the inverse DFT. The expander is Due to its mathematical tractability, the Mean Square based on numerical methods: we do one iteration in a fixed- Error (MSE) is the elected choice for the distortion measure point equation, and then we fine-tune the result using Broy- in many coding applications. In audio coding, however, the den’s method. -
Audio Coding for Digital Broadcasting
Recommendation ITU-R BS.1196-7 (01/2019) Audio coding for digital broadcasting BS Series Broadcasting service (sound) ii Rec. ITU-R BS.1196-7 Foreword The role of the Radiocommunication Sector is to ensure the rational, equitable, efficient and economical use of the radio- frequency spectrum by all radiocommunication services, including satellite services, and carry out studies without limit of frequency range on the basis of which Recommendations are adopted. The regulatory and policy functions of the Radiocommunication Sector are performed by World and Regional Radiocommunication Conferences and Radiocommunication Assemblies supported by Study Groups. Policy on Intellectual Property Right (IPR) ITU-R policy on IPR is described in the Common Patent Policy for ITU-T/ITU-R/ISO/IEC referenced in Resolution ITU-R 1. Forms to be used for the submission of patent statements and licensing declarations by patent holders are available from http://www.itu.int/ITU-R/go/patents/en where the Guidelines for Implementation of the Common Patent Policy for ITU-T/ITU-R/ISO/IEC and the ITU-R patent information database can also be found. Series of ITU-R Recommendations (Also available online at http://www.itu.int/publ/R-REC/en) Series Title BO Satellite delivery BR Recording for production, archival and play-out; film for television BS Broadcasting service (sound) BT Broadcasting service (television) F Fixed service M Mobile, radiodetermination, amateur and related satellite services P Radiowave propagation RA Radio astronomy RS Remote sensing systems S Fixed-satellite service SA Space applications and meteorology SF Frequency sharing and coordination between fixed-satellite and fixed service systems SM Spectrum management SNG Satellite news gathering TF Time signals and frequency standards emissions V Vocabulary and related subjects Note: This ITU-R Recommendation was approved in English under the procedure detailed in Resolution ITU-R 1. -
Challenges in Relaying Video Back to Mission Control
Challenges in Relaying Video Back to Mission Control CONTENTS 1 Introduction ..................................................................................................................................................................................... 2 2 Encoding ........................................................................................................................................................................................... 3 3 Time is of the Essence .............................................................................................................................................................. 3 4 The Reality ....................................................................................................................................................................................... 5 5 The Tradeoff .................................................................................................................................................................................... 5 6 Alleviations and Implementation ......................................................................................................................................... 6 7 Conclusion ....................................................................................................................................................................................... 7 White Paper Revision 2 July 2020 By Christopher Fadeley Using a customizable hardware-accelerated encoder is essential to delivering the high -
(A/V Codecs) REDCODE RAW (.R3D) ARRIRAW
What is a Codec? Codec is a portmanteau of either "Compressor-Decompressor" or "Coder-Decoder," which describes a device or program capable of performing transformations on a data stream or signal. Codecs encode a stream or signal for transmission, storage or encryption and decode it for viewing or editing. Codecs are often used in videoconferencing and streaming media solutions. A video codec converts analog video signals from a video camera into digital signals for transmission. It then converts the digital signals back to analog for display. An audio codec converts analog audio signals from a microphone into digital signals for transmission. It then converts the digital signals back to analog for playing. The raw encoded form of audio and video data is often called essence, to distinguish it from the metadata information that together make up the information content of the stream and any "wrapper" data that is then added to aid access to or improve the robustness of the stream. Most codecs are lossy, in order to get a reasonably small file size. There are lossless codecs as well, but for most purposes the almost imperceptible increase in quality is not worth the considerable increase in data size. The main exception is if the data will undergo more processing in the future, in which case the repeated lossy encoding would damage the eventual quality too much. Many multimedia data streams need to contain both audio and video data, and often some form of metadata that permits synchronization of the audio and video. Each of these three streams may be handled by different programs, processes, or hardware; but for the multimedia data stream to be useful in stored or transmitted form, they must be encapsulated together in a container format. -
MC14SM5567 PCM Codec-Filter
Product Preview Freescale Semiconductor, Inc. MC14SM5567/D Rev. 0, 4/2002 MC14SM5567 PCM Codec-Filter The MC14SM5567 is a per channel PCM Codec-Filter, designed to operate in both synchronous and asynchronous applications. This device 20 performs the voice digitization and reconstruction as well as the band 1 limiting and smoothing required for (A-Law) PCM systems. DW SUFFIX This device has an input operational amplifier whose output is the input SOG PACKAGE CASE 751D to the encoder section. The encoder section immediately low-pass filters the analog signal with an RC filter to eliminate very-high-frequency noise from being modulated down to the pass band by the switched capacitor filter. From the active R-C filter, the analog signal is converted to a differential signal. From this point, all analog signal processing is done differentially. This allows processing of an analog signal that is twice the amplitude allowed by a single-ended design, which reduces the significance of noise to both the inverted and non-inverted signal paths. Another advantage of this differential design is that noise injected via the power supplies is a common mode signal that is cancelled when the inverted and non-inverted signals are recombined. This dramatically improves the power supply rejection ratio. After the differential converter, a differential switched capacitor filter band passes the analog signal from 200 Hz to 3400 Hz before the signal is digitized by the differential compressing A/D converter. The decoder accepts PCM data and expands it using a differential D/A converter. The output of the D/A is low-pass filtered at 3400 Hz and sinX/X compensated by a differential switched capacitor filter. -
Hikvision H.264+ Encoding Technology
WHITE PAPER Hikvision H.264+ Encoding Technology Encoding Improvement / Higher Transmission / Efficiency Storage Savings 2 Contents 1. Introduction .............................................................................................. 3 2. Background ............................................................................................... 3 3. Key Technologies .................................................................................... 4 3.1 Predictive Encoding ........................................................................ 4 3.2 Noise Suppression.......................................................................... 8 3.3 Long-Term Bitrate Control........................................................... 9 4. Applications ............................................................................................ 11 5. Conclusion............................................................................................... 11 Hikvision H.264+ Encoding Technology 3 1. INTRODUCTION As the global market leader in video surveillance products, Hikvision Digital Technology Co., Ltd., continues to strive for enhancement of its products through application of the latest in technology. H.264+ Advanced Video Coding (AVC) optimizes compression beyond the current H.264 standard. Through the combination of intelligent analysis technology with predictive encoding, noise suppression, and long-term bitrate control, Hikvision is meeting the demand for higher resolution at reduced bandwidths. Our customers will benefit -
Subjective Assessment of Audio Quality – the Means and Methods Within the EBU
Subjective assessment of audio quality – the means and methods within the EBU W. Hoeg (Deutsch Telekom Berkom) L. Christensen (Danmarks Radio) R. Walker (BBC) This article presents a number of useful means and methods for the subjective quality assessment of audio programme material in radio and television, developed and verified by EBU Project Group, P/LIST. The methods defined in several new 1. Introduction EBU Recommendations and Technical Documents are suitable for The existing EBU Recommendation, R22 [1], both operational and training states that “the amount of sound programme ma- purposes in broadcasting terial which is exchanged between EBU Members, organizations. and between EBU Members and other production organizations, continues to increase” and that “the only sufficient method of assessing the bal- ance of features which contribute to the quality of An essential prerequisite for ensuring a uniform the sound in a programme is by listening to it.” high quality of sound programmes is to standard- ize the means and methods required for their as- Therefore, “listening” is an integral part of all sessment. The subjective assessment of sound sound and television programme-making opera- quality has for a long time been carried out by tions. Despite the very significant advances of international organizations such as the (former) modern sound monitoring and measurement OIRT [2][3][4], the (former) CCIR (now ITU-R) technology, these essentially objective solutions [5] and the Member organizations of the EBU remain unable to tell us what the programme will itself. It became increasingly important that com- really sound like to the listener at home. -
Optimized Bitrate Ladders for Adaptive Video Streaming with Deep Reinforcement Learning
Optimized Bitrate Ladders for Adaptive Video Streaming with Deep Reinforcement Learning ∗ Tianchi Huang1, Lifeng Sun1,2,3 1Dept. of CS & Tech., 2BNRist, Tsinghua University. 3Key Laboratory of Pervasive Computing, China ABSTRACT Transcoding Online Stage Video Quality Stage In the adaptive video streaming scenario, videos are pre-chunked Storage Cost and pre-encoded according to a set of resolution-bitrate/quality Deploy pairs on the server-side, namely bitrate ladder. Hence, we pro- … … pose DeepLadder, which adopts state-of-the-art deep reinforcement learning (DRL) method to optimize the bitrate ladder by consid- Transcoding Server ering video content features, current network capacities, as well Raw Videos Video Chunks NN-based as the storage cost. Experimental results on both Constant Bi- Decison trate (CBR) and Variable Bitrate (VBR)-encoded videos demonstrate Network & ABR Status Feedback that DeepLadder significantly improvements on average video qual- ity, bandwidth utilization, and storage overhead in comparison to Figure 1: An Overview of DeepLadder’s System. We leverage prior work. a NN-based decision model for constructing the proper bi- trate ladders, and transcode the video according to the as- CCS CONCEPTS signed settings. • Information systems → Multimedia streaming; • Computing and solve the problem mathematically. In this poster, we propose methodologies → Neural networks; DeepLadder, a per-chunk video transcoding system. Technically, we set video contents, current network traffic distributions, past KEYWORDS actions as the state, and utilize a neural network (NN) to deter- Bitrate Ladder Optimization, Deep Reinforcement Learning. mine the proper action for each resolution autoregressively. Unlike the traditional bitrate ladder method that outputs all candidates ACM Reference Format: at one step, we model the optimization process as a Markov Deci- Tianchi Huang, Lifeng Sun. -
TMS320C6000 U-Law and A-Law Companding with Software Or The
Application Report SPRA634 - April 2000 TMS320C6000t µ-Law and A-Law Companding with Software or the McBSP Mark A. Castellano Digital Signal Processing Solutions Todd Hiers Rebecca Ma ABSTRACT This document describes how to perform data companding with the TMS320C6000 digital signal processors (DSP). Companding refers to the compression and expansion of transfer data before and after transmission, respectively. The multichannel buffered serial port (McBSP) in the TMS320C6000 supports two companding formats: µ-Law and A-Law. Both companding formats are specified in the CCITT G.711 recommendation [1]. This application report discusses how to use the McBSP to perform data companding in both the µ-Law and A-Law formats. This document also discusses how to perform companding of data not coming into the device but instead located in a buffer in processor memory. Sample McBSP setup codes are provided for reference. In addition, this application report discusses µ-Law and A-Law companding in software. The appendix provides software companding codes and a brief discussion on the companding theory. Contents 1 Design Problem . 2 2 Overview . 2 3 Companding With the McBSP. 3 3.1 µ-Law Companding . 3 3.1.1 McBSP Register Configuration for µ-Law Companding. 4 3.2 A-Law Companding. 4 3.2.1 McBSP Register Configuration for A-Law Companding. 4 3.3 Companding Internal Data. 5 3.3.1 Non-DLB Mode. 5 3.3.2 DLB Mode . 6 3.4 Sample C Functions. 6 4 Companding With Software. 6 5 Conclusion . 7 6 References . 7 TMS320C6000 is a trademark of Texas Instruments Incorporated. -
Equalization
CHAPTER 3 Equalization 35 The most intuitive effect — OK EQ is EZ U need a Hi EQ IQ MIX SMART QUICK START: Equalization GOALS ■ Fix spectral problems such as rumble, hum and buzz, pops and wind, proximity, and hiss. ■ Fit things into the mix by leveraging of any spectral openings available and through complementary cuts and boosts on spectrally competitive tracks. ■ Feature those aspects of each instrument that players and music fans like most. GEAR ■ Master the user controls for parametric, semiparametric, program, graphic, shelving, high-pass and low-pass filters. ■ Choose the equalizer with the capabilities you need, focusing particularly on the parameters, slopes, and number of bands available. Home stereos have tone controls. We in the studio get equalizers. Perhaps bet- ter described as a “spectral modifier” or “frequency-specific amplitude adjuster,” the equalizer allows the mix engineer to increase or decrease the level of specific frequency ranges within a signal. Having an equalizer is like having multiple volume knobs for a single track. Unlike the volume knob that attenuates or boosts an entire signal, the equalizer is the tool used to turn down or turn up specific frequency portions of any audio track . Out of all signal-processing tools that we use while mixing and tracking, EQ is probably the easiest, most intuitive to use—at first. Advanced applications of equalization, however, demand a deep understanding of the effect and all its Mix Smart. © 2011 Elsevier Inc. All rights reserved. 36 Mix Smart possibilities. Don't underestimate the intellectual challenge and creative poten- tial of this essential mix processor. -
Cube Encoder and Decoder Reference Guide
CUBE ENCODER AND DECODER REFERENCE GUIDE © 2018 Teradek, LLC. All Rights Reserved. TABLE OF CONTENTS 1. Introduction ................................................................................ 3 Support Resources ........................................................ 3 Disclaimer ......................................................................... 3 Warning ............................................................................. 3 HEVC Products ............................................................... 3 HEVC Content ................................................................. 3 Physical Properties ........................................................ 4 2. Getting Started .......................................................................... 5 Power Your Device ......................................................... 5 Connect to a Network .................................................. 6 Choose Your Application .............................................. 7 Choose a Stream Mode ............................................... 9 3. Encoder Configuration ..........................................................10 Video/Audio Input .......................................................12 Color Management ......................................................13 Encoder ...........................................................................14 Network Interfaces .....................................................15 Cloud Services ..............................................................17 -
4. MPEG Layer-3 Audio Encoding
MPEG Layer-3 An introduction to MPEG Layer-3 K. Brandenburg and H. Popp Fraunhofer Institut für Integrierte Schaltungen (IIS) MPEG Layer-3, otherwise known as MP3, has generated a phenomenal interest among Internet users, or at least among those who want to download highly-compressed digital audio files at near-CD quality. This article provides an introduction to the work of the MPEG group which was, and still is, responsible for bringing this open (i.e. non-proprietary) compression standard to the forefront of Internet audio downloads. 1. Introduction The audio coding scheme MPEG Layer-3 will soon celebrate its 10th birthday, having been standardized in 1991. In its first years, the scheme was mainly used within DSP- based codecs for studio applications, allowing professionals to use ISDN phone lines as cost-effective music links with high sound quality. In 1995, MPEG Layer-3 was selected as the audio format for the digital satellite broadcasting system developed by World- Space. This was its first step into the mass market. Its second step soon followed, due to the use of the Internet for the electronic distribution of music. Here, the proliferation of audio material – coded with MPEG Layer-3 (aka MP3) – has shown an exponential growth since 1995. By early 1999, “.mp3” had become the most popular search term on the Web (according to http://www.searchterms.com). In 1998, the “MPMAN” (by Saehan Information Systems, South Korea) was the first portable MP3 player, pioneering the road for numerous other manufacturers of consumer electronics. “MP3” has been featured in many articles, mostly on the business pages of newspapers and periodicals, due to its enormous impact on the recording industry.