Video Coding with Superimposed Motion-Compensated Signals

Total Page:16

File Type:pdf, Size:1020Kb

Video Coding with Superimposed Motion-Compensated Signals Video Coding with Superimposed Motion-Compensated Signals Videocodierung mituberlagerten ¨ bewegungskompensierten Signalen Der Technischen Fakult¨at der Friedrich-Alexander-Universit¨at Erlangen-N¨urnberg zur Erlangung des Grades Doktor-Ingenieur vorgelegt von Markus Helmut Flierl Erlangen, 2003 Als Dissertation genehmigt von der Technischen Fakult¨at der Friedrich-Alexander-Universit¨at Erlangen-N¨urnberg Tag der Einreichung: 13.03.2003 Tag der Promotion: 27.06.2003 Dekan: Prof. Dr. rer. Nat. A. Winnacker Berichterstatter: Prof. Dr.-Ing. B. Girod Prof. Dr.-Ing. A. Kaup Acknowledgments This work required a substantial effort and I am very grateful to many people who made this endeavor possible. I would like to thank my supervisor Prof. Bernd Girod for the opportunity to join his group in Erlangen and Stanford as well as to benefit from the inspiring environment. I would also like to thank Prof. Andr´e Kaup for his interest in my work and reviewing my thesis as well as Prof. Heinrich Niemann and Prof. Wolfgang Koch for joining the committee. I am thankful to my fellow students Anne Aaron, Chuo-Ling Chang, Jacob Charareski, Joachim Eggers, Peter Eisert, Niko F¨arber, Sang-Eun Han, Frank Hartung, Mark Kalman, Yi Liang, Marcus Magnor, James Mammen, Prashant Ramanathan, Shantanu Rane, Marion Schabert, Eric Setton, Wolfgang S¨orgel, Eckehard Steinbach, Klaus Stuhlm¨uller, Jonathan Su, Matthias Teschner, Lutz Trautmann, Thomas Wiegand, and Rui Zhang for many stimulating discussions, joint work, and proofreading. In addition, I am thankful for helpful discussions with Dr. Nick Kingsbury. I am also grateful to the people at the Telecommunications Laboratory and Information Systems Laboratory, especially to Ursula Arnold and Kelly Yilmaz for their invaluable adminis- trative support. My special thanks belong to my family and friends for all the support they gave. Erkenne die Musik dessen vollendete T¨one verklingen. F¨uhle die Stille die zum Tr¨aumen verf¨uhrt. Erfahre den unendlichen Raum der hinschwindenden Ansammlung sich vollendender T¨one. — Deinem Odysseus vii Contents 1 Introduction 1 2 Background and Related Work 7 2.1CodingofVideoSignals............................ 7 2.2MotionCompensation............................. 8 2.2.1 BidirectionalMotionCompensation.................. 9 2.2.2 Overlapped Block Motion Compensation . 10 2.2.3 Variable Block Size Motion Compensation . 11 2.2.4 Multiframe Motion Compensation . 12 2.2.5 Superimposed Motion Compensation . 13 2.3MotionEstimation............................... 16 2.3.1 Rate-Constrained Motion Estimation . 16 2.3.2 Rate-Constrained Estimation of Superimposed Motion . 18 2.3.3 QuantizerSelectionattheResidualEncoder............. 20 2.3.4 EfficientMotionEstimation...................... 22 2.4 Theory of Motion-Compensated Prediction . 23 2.4.1 FrameSignalModel.......................... 24 2.4.2 Signal Model for Motion-Compensated Prediction . 25 2.4.3 Signal Model for Multihypothesis Prediction . 26 2.4.4 PerformanceMeasures......................... 28 2.4.5 Conclusions............................... 29 2.5 Three-Dimensional Subband Coding of Video . 29 2.5.1 Motion Compensation and Subband Coding . 30 2.5.2 Motion-CompensatedLiftedWavelets................. 31 viii Contents 3 Motion-Compensated Prediction with Complementary Hypotheses 33 3.1Introduction................................... 33 3.2 Extended Model for Superimposed Motion-Compensated Prediction . 34 3.2.1 Superimposed Prediction and Correlated Displacement Error . 34 3.2.2 ComplementaryHypotheses...................... 36 3.2.3 GradientofthePredictionErrorVariance.............. 41 3.3HypotheseswithAdditiveNoise........................ 42 3.3.1 AveragingFilter............................. 43 3.3.2 WienerFilter.............................. 45 3.4Forward-AdaptiveHypothesisSwitching................... 52 3.4.1 SignalModelforHypothesisSwitching................ 53 3.4.2 MinimizingtheRadialDisplacementError.............. 53 3.4.3 EquivalentPredictor.......................... 56 3.4.4 Motion Compensation with Complementary Hypotheses and For- ward-AdaptiveHypothesisSwitching................. 57 3.5PictureswithVaryingNumberofHypotheses................ 59 3.6Conclusions................................... 61 4 ITU-T Rec. H.263 and Superimposed Prediction 63 4.1Introduction................................... 63 4.2VideoCodingwithSuperimposedMotion................... 64 4.2.1 SyntaxExtensions........................... 64 4.2.2 CoderControl.............................. 64 4.3ExperimentalResults.............................. 66 4.3.1 Multiple Hypotheses for Constant Block Size . 66 4.3.2 Multiple Hypotheses for Variable Block Size . 67 4.3.3 Multiple Hypotheses and Multiple Reference Pictures . 69 4.4Conclusions................................... 76 5 ITU-T Rec. H.264 and Generalized B-Pictures 77 5.1Introduction................................... 77 5.2PredictionModesforB-Pictures........................ 78 5.2.1 Overview................................ 78 5.2.2 DirectMode............................... 79 5.2.3 SuperpositionMode.......................... 80 5.2.4 Rate-Distortion Performance of Individual Modes . 80 5.3SuperimposedPrediction............................ 83 5.3.1 Bidirectionalvs.SuperpositionMode................. 83 Contents ix 5.3.2 Two Combined Forward Prediction Signals . 84 5.3.3 EntropyCoding............................. 89 5.4EncoderIssues................................. 90 5.4.1 RateConstrainedModeDecision................... 90 5.4.2 Rate Constrained Motion Estimation . 91 5.4.3 Rate Constrained Superimposed Motion Estimation . 92 5.4.4 Improving Overall Rate-Distortion Performance . 93 5.5Conclusions................................... 97 6 Motion Compensation for Groups of Pictures 99 6.1Introduction................................... 99 6.2CodingScheme................................. 99 6.2.1 Motion-Compensated Lifted Haar Wavelet . 100 6.2.2 Motion-Compensated Lifted 5/3 Wavelet . 101 6.2.3 ExperimentalResults..........................101 6.3 A Mathematical Model of Motion-Compensated Subband Coding . 108 6.3.1 Motion-Compensated Lifted Haar Wavelet . 108 6.3.2 Motion-Compensated Lifted 5/3 Wavelet . 109 6.3.3 SignalModel..............................109 6.3.4 TransformCodingGain........................113 6.4Conclusions...................................115 7 Summary 117 A Mathematical Results 123 A.1 Singularities of the Displacement Error Covariance Matrix . 123 A.2AClassofMatricesandtheirEigenvalues..................123 A.3 Inverse of the Power Spectral Density Matrix . 124 A.4PowerSpectralDensityofaFrame......................126 Glossary 127 Bibliography 131 x Contents Inhaltsverzeichnis 1 1Einf¨uhrung 1 2 Grundlagen und Stand der Forschung 7 2.1CodierungvonVideosignalen......................... 7 2.2Bewegungskompensation............................ 8 2.2.1 BidirektionaleBewegungskompensation................ 9 2.2.2 Bewegungskompensation mituberlappenden ¨ Bl¨ocken........ 10 2.2.3 Bewegungskompensation mit Bl¨ocken variabler Gr¨oße........ 11 2.2.4 Bewegungskompensation mit mehreren Referenzbildern . 12 2.2.5 UberlagerteBewegungskompensation.................¨ 13 2.3 Bewegungssch¨atzung.............................. 16 2.3.1 Bewegungssch¨atzungmitRatennebenbedingung........... 16 2.3.2 Sch¨atzunguberlagerter ¨ Bewegung mit Ratennebenbedingung . 18 2.3.3 Wahl des Quantisierers beim Pr¨adiktionsfehler-Encoder....... 20 2.3.4 Effiziente Bewegungssch¨atzung..................... 22 2.4 Theorie der bewegungskompensierten Pr¨adiktion............... 23 2.4.1 Bildsignalmodell............................ 24 2.4.2 Signalmodell f¨ur bewegungskompensierte Pr¨adiktion......... 25 2.4.3 Signalmodell f¨ur Multihypothesen-Pr¨adiktion............ 26 2.4.4 Bewertungsmaße............................ 28 2.4.5 Schlussfolgerungen........................... 29 2.5Drei-DimensionaleTeilbandcodierungvonVideo............... 29 2.5.1 Bewegungskompensation und Teilbandcodierung . 30 2.5.2 BewegungskompensierteWavelets................... 31 3 Bewegungskompensierte Pr¨adiktion mit komplement¨aren Hypothesen 33 3.1Einleitung.................................... 33 3.2 Erweitertes Modell f¨ur dieuberlagerte ¨ bewegungskompensierte Pr¨adiktion . 34 3.2.1 Uberlagerte¨ Pr¨adiktion und korrelierte Versatzfehler . 34 3.2.2 Komplement¨areHypothesen...................... 36 3.2.3 Gradient der Pr¨adiktionsfehlervarianz................. 41 3.3HypothesenmitadditivemRauschen..................... 42 3.3.1 Mittelungsfilter............................. 43 3.3.2 WienerFilter.............................. 45 1This translation of the table of contents is a requirement according to the ”Pr¨ufungsordnung f¨ur die Technischen Fakult¨at der Friedrich-Alexander-Universit¨at Erlangen-N¨urnberg”. Contents xi 3.4AdaptiveHypothesenauswahl......................... 52 3.4.1 Signalmodell f¨urdieAuswahlvonHypothesen............ 53 3.4.2 Minimierung des radialen Versatzfehlers . 53 3.4.3 Aquivalenter¨ Pr¨adiktor......................... 56 3.4.4 Bewegungskompensation mit komplement¨aren und adaptiv aus- gew¨ahltenHypothesen......................... 57 3.5BildermitvariierenderAnzahlanHypothesen................ 59 3.6Schlussfolgerungen............................... 61 4 ITU-T Rec. H.263 unduberlagerte ¨ Pr¨adiktion 63 4.1Einleitung.................................... 63 4.2 Videocodierung mituberlagerterBewegung.................. ¨ 64 4.2.1 SyntaxErweiterung........................... 64 4.2.2 Algorithmen f¨urdenEncoder..................... 64 4.3ExperimentelleErgebnisse..........................
Recommended publications
  • LOW COMPLEXITY H.264 to VC-1 TRANSCODER by VIDHYA
    LOW COMPLEXITY H.264 TO VC-1 TRANSCODER by VIDHYA VIJAYAKUMAR Presented to the Faculty of the Graduate School of The University of Texas at Arlington in Partial Fulfillment of the Requirements for the Degree of MASTER OF SCIENCE IN ELECTRICAL ENGINEERING THE UNIVERSITY OF TEXAS AT ARLINGTON AUGUST 2010 Copyright © by Vidhya Vijayakumar 2010 All Rights Reserved ACKNOWLEDGEMENTS As true as it would be with any research effort, this endeavor would not have been possible without the guidance and support of a number of people whom I stand to thank at this juncture. First and foremost, I express my sincere gratitude to my advisor and mentor, Dr. K.R. Rao, who has been the backbone of this whole exercise. I am greatly indebted for all the things that I have learnt from him, academically and otherwise. I thank Dr. Ishfaq Ahmad for being my co-advisor and mentor and for his invaluable guidance and support. I was fortunate to work with Dr. Ahmad as his research assistant on the latest trends in video compression and it has been an invaluable experience. I thank my mentor, Mr. Vishy Swaminathan, and my team members at Adobe Systems for giving me an opportunity to work in the industry and guide me during my internship. I would like to thank the other members of my advisory committee Dr. W. Alan Davis and Dr. William E Dillon for reviewing the thesis document and offering insightful comments. I express my gratitude Dr. Jonathan Bredow and the Electrical Engineering department for purchasing the software required for this thesis and giving me the chance to work on cutting edge technologies.
    [Show full text]
  • Video Coding Standards
    Module 8 Video Coding Standards Version 2 ECE IIT, Kharagpur Lesson 23 MPEG-1 standards Version 2 ECE IIT, Kharagpur Lesson objectives At the end of this lesson, the students should be able to : 1. Enlist the major video coding standards 2. State the basic objectives of MPEG-1 standard. 3. Enlist the set of constrained parameters in MPEG-1 4. Define the I- P- and B-pictures 5. Present the hierarchical data structure of MPEG-1 6. Define the macroblock modes supported by MPEG-1 23.0 Introduction In lesson 21 and lesson 22, we studied how to perform motion estimation and thereby temporally predict the video frames to exploit significant temporal redundancies present in the video sequence. The error in temporal prediction is encoded by standard transform domain techniques like the DCT, followed by quantization and entropy coding to exploit the spatial and statistical redundancies and achieve significant video compression. The video codecs therefore follow a hybrid coding structure in which DPCM is adopted in temporal domain and DCT or other transform domain techniques in spatial domain. Efforts to standardize video data exchange via storage media or via communication networks are actively in progress since early 1980s. A number of international video and audio standardization activities started within the International Telephone Consultative Committee (CCITT), followed by the International Radio Consultative Committee (CCIR), and the International Standards Organization / International Electrotechnical Commission (ISO/IEC). An experts group, known as the Motion Pictures Expects Group (MPEG) was established in 1988 in the framework of the Joint ISO/IEC Technical Committee with an objective to develop standards for coded representation of moving pictures, associated audio, and their combination for storage and retrieval of digital media.
    [Show full text]
  • AVC to the Max: How to Configure Encoder
    Contents Company overview …. ………………………………………………………………… 3 Introduction…………………………………………………………………………… 4 What is AVC….………………………………………………………………………… 6 Making sense of profiles, levels, and bitrate………………………………………... 7 Group of pictures and its structure..………………………………………………… 11 Macroblocks: partitioning and prediction modes….………………………………. 14 Eliminating spatial redundancy……………………………………………………… 15 Eliminating temporal redundancy……...……………………………………………. 17 Adaptive quantization……...………………………………………………………… 24 Deblocking filtering….….…………………………………………………………….. 26 Entropy encoding…………………………………….……………………………….. 2 8 Conclusion…………………………………………………………………………….. 29 Contact details..………………………………………………………………………. 30 2 www.elecard.com Company overview Elecard company, founded in 1988, is a leading provider of software products for encoding, decoding, processing, monitoring and analysis of video and audio data in 9700 companies various formats. Elecard is a vendor of professional software products and software development kits (SDKs); products for in - depth high - quality analysis and monitoring of the media content; countries 1 50 solutions for IPTV and OTT projects, digital TV broadcasting and video streaming; transcoding servers. Elecard is based in the United States, Russia, and China with 20M users headquarters located in Tomsk, Russia. Elecard products are highly appreciated and widely used by the leaders of IT industry such as Intel, Cisco, Netflix, Huawei, Blackmagic Design, etc. For more information, please visit www.elecard.com. 3 www.elecard.com Introduction Video compression is the key step in video processing. Compression allows broadcasters and premium TV providers to deliver their content to their audience. Many video compression standards currently exist in TV broadcasting. Each standard has different properties, some of which are better suited to traditional live TV while others are more suited to video on demand (VoD). Two basic standards can be identified in the history of video compression: • MPEG-2, a legacy codec used for SD video and early digital broadcasting.
    [Show full text]
  • Video Coding Standards 1 Videovideo Codingcoding Standardsstandards
    VideoVideo CodingCoding StandardsStandards • H.120 • H.261 • MPEG-1 and MPEG-2/H.262 • H.263 • MPEG-4 Thomas Wiegand: Digital Image Communication Video Coding Standards 1 VideoVideo CodingCoding StandardsStandards MPEG-2 digital TV 2 -6 Mbps ITU-R 601 166 Mbit/s H.261 ISDN 64 kbps Picture phone H.263 PSTN < 28.8 kbps picture phone Thomas Wiegand: Digital Image Communication Video Coding Standards 2 H.120:H.120: TheThe FirstFirst DigitalDigital VideoVideo CodingCoding StandardStandard • ITU-T (ex-CCITT) Rec. H.120: The first digital video coding standard (1984) • v1 (1984) had conditional replenishment, DPCM, scalar quantization, variable-length coding, switch for quincunx sampling • v2 (1988) added motion compensation and background prediction • Operated at 1544 (NTSC) and 2048 (PAL) kbps • Few units made, essentially not in use today Thomas Wiegand: Digital Image Communication Video Coding Standards 3 H.261:H.261: TheThe BasisBasis ofof ModernModern VideoVideo CompressionCompression • ITU-T (ex-CCITT) Rec. H.261: The first widespread practical success • First design (late ’80s) embodying typical structure that dominates today: 16x16 macroblock motion compensation, 8x8 DCT, scalar quantization, and variable-length coding • Other key aspects: loop filter, integer-pel motion compensation accuracy, 2-D VLC for coefficients • Operated at 64-2048 kbps • Still in use, although mostly as a backward- compatibility feature – overtaken by H.263 Thomas Wiegand: Digital Image Communication Video Coding Standards 4 H.261&3H.261&3 MacroblockMacroblock
    [Show full text]
  • Motion Compensation on DCT Domain
    EURASIP Journal on Applied Signal Processing 2001:3, 147–162 © 2001 Hindawi Publishing Corporation Motion Compensation on DCT Domain Ut-Va Koc Lucent Technologies Bell Labs, 600 Mountain Avenue, Murray Hill, NJ 07974, USA Email: [email protected] K. J. Ray Liu Department of Electrical and Computer Engineering and Institute for Systems Research, University of Maryland, College Park, MD 20742, USA Email: [email protected] Received 21 May 2001 and in revised form 21 September 2001 Alternative fully DCT-based video codec architectures have been proposed in the past to address the shortcomings of the conven- tional hybrid motion compensated DCT video codec structures traditionally chosen as the basis of implementation of standard- compliant codecs. However, no prior effort has been made to ensure interoperability of these two drastically different architectures so that fully DCT-based video codecs are fully compatible with the existing video coding standards. In this paper, we establish the criteria for matching conventional codecs with fully DCT-based codecs. We find that the key to this interoperability lies in the heart of the implementation of motion compensation modules performed in the spatial and transform domains at both the encoder and the decoder. Specifically,if the spatial-domain motion compensation is compatible with the transform-domain motion compensation, then the states in both the coder and the decoder will keep track of each other even after a long series of P-frames. Otherwise, the states will diverge in proportion to the number of P-frames between two I-frames. This sets an important criterion for the development of any DCT-based motion compensation schemes.
    [Show full text]
  • White Paper Version of July 2015
    White paper Digital Video File Recommendation Version of July 2015 edited by Whitepaper 1 DIGITAL VIDEO FILE RECOMMENDATIONS ............................ 4 1.1 Preamble .........................................................................................................................4 1.2 Scope ...............................................................................................................................5 1.3 File Formats ....................................................................................................................5 1.4 Codecs ............................................................................................................................5 1.4.1 Browsing .................................................................................................... 5 1.4.2 Acquisition ................................................................................................. 6 1.4.3 Programme Contribution ........................................................................... 7 1.4.4 Postproduction .......................................................................................... 7 1.4.5 Broadcast ................................................................................................... 7 1.4.6 News & Magazines & Sports ..................................................................... 8 1.4.7 High Definition ........................................................................................... 8 1.5 General Requirements ..................................................................................................9
    [Show full text]
  • Axis Communications – Have Experienced Increasing Overlap in Their Standardization Efforts
    WHITE PAPER An explanation of video compression techniques. Table of contents 1. Introduction to compression techniques 4 2. Standardization organizations 4 3. Two basic standards: JPEG and MPEG 4 4. The next step: H.264 5 5. The basics of compression 5 5.1 Lossless and lossy compression 6 5.2 Latency 6 6. An overview of compression formats 6 6.1 JPEG 6 6.2 Motion JPEG 7 6.3 JPEG 2000 7 6.4 Motion JPEG 2000 7 6.5 H.261/H.263 8 6.6 MPEG-1 8 6.7 MPEG-2 9 6.8 MPEG-3 9 6.9 MPEG-4 9 6.10 H.264 10 6.11 MPEG-7 10 6.12 MPEG-21 10 7. More on MPEG compression 10 7.1 Frame types 11 7.2 Group of Pictures 11 7.3 Variable and constant bit rates 12 25. MPEG comparison 12 26. Conclusion – Still pictures 13 27. Conclusion – Motion pictures 13 28. Acronyms 15 1. Introduction to compression techniques JPEG, Motion JPEG and MPEG are three well-used acronyms used to describe different types of image compression format. But what do they mean, and why are they so relevant to today’s rapidly expanding surveillance market? This White Paper describes the differences, and aims to provide a few answers as to why they are so important and for which surveillance applications they are suitable. When designing a networked digital surveillance application, developers need to initially consider the following factors: > Is a still picture or a video sequence required? > What is the available network bandwidth? > What image degradation is allowed due to compression – so called artifacts? > What is the budget for the system? When an ordinary analog video sequence is digitized according to the standard CCIR 601, it can consume as much as 165 Mbps, which is 165 million bits every second.
    [Show full text]
  • MPEG-2 Fundamentals for Broadcast and Post-Production Engineers
    Mpeg2 Page 1 of 21 MPEG-2 Fundamentals for Broadcast and Post-Production Engineers A Video and Networking Division White Paper By Adolfo Rodriguez, David K. Fibush and Steven C. Bilow July, 1996 INTRODUCTION This paper will discuss the digital representation of video as defined in the MPEG-2 standard and will examine some crucial, strategic issues regarding the newly adopted 4:2:2 profile at main level. Specifically, we will cover the rationale for this new profile and will illustrate the benefits it provides to the broadcast and post-production communities. The intent of this paper is to articulate the appropriate applications of the various MPEG profiles and levels, and to demonstrate the fundamental necessity for the new 4:2:2 profile. We will also address the encoder/decoder compatibility required by the standard, the pervasive need for a 4:2:2 video representation, and the method by which the 4:2:2 profile at main level allows MPEG-2 to provide these advantages in a cost-effective manner. The new MPEG 4:2:2 profile at main level has recently become an official part of the MPEG standard. To broadcasters and production houses that do not accept the low chroma bandwidth options provided by previous MPEG main level profiles, the 4:2:2 profile at main level will prove to be the long-awaited solution for contribution quality MPEG compressed video. The Tektronix perspective is the result of a 50-year history of instrumentation, broadcast, and video equipment design. It is also the product of our success in helping the MPEG-2 standards body toward the provision of this new, cost-effective support for broadcast-quality video.
    [Show full text]
  • Multiview Video Coding Extension of the H.264/AVC Standard
    Multiview Video Coding Extension of the H.264/AVC Standard Ognjen Nemþiü 1, Snježana Rimac-Drlje 2, Mario Vranješ 2 1 Supra Net Projekt d.o.o., Zagreb, Croatia, [email protected] 2 University of Osijek, Faculty of Electrical Engineering, Osijek, Croatia, [email protected] Abstract - An overview of the new Multiview Video Coding approved an amendment of the ITU-T Rec. H.264 & ISO/IEC (MVC) extension of the H.264/AVC standard is given in this 14996-10 Advanced Video Coding (AVC) standard on paper. The benefits of multiview video coding are examined by Multiview Video Coding. encoding two data sets captured with multiple cameras. These sequences were encoded using MVC reference software and the results are compared with references obtained by encoding II. MULTIVIEW VIDEO CODING multiple views independently with H.264/AVC reference MVC was designed as an amendment of the H.264/AVC software. Experimental quality comparison of coding efficiency is and uses improved coding tools which have made H.264/AVC given. The peak signal-to-noise ration (PSNR) and Structural so superior over older standards. It is based on conventional Similarity (SSIM) Index are used as objective quality metrics. block-based motion-compensated video coding, with several new features which significantly improve its rate-distortion Keywords - H.264/AVC, MVC, inter-view prediction, temporal performance. The main new features of H.264/AVC, also used prediction, video quality by MVC, are the following: • I. INTRODUCTION variable block-size motion-compensated prediction with the block size down to 4x4 pixels; The H.264/AVC is the state-of-the-art video coding standard, [1], developed jointly by the ITU-T Video Coding • quarter-pixel motion vector accuracy; Experts Group (VCEG) and the ISO/IEC Moving Pictures • Experts Group (MPEG).
    [Show full text]
  • Fundamentals of Digital Video Compression
    FUNDAMENTALS OF DIGITAL VIDEO COMPRESSION MPEG Compression Streaming Video and Progressive Download © 2016 Pearson Education, Inc., Hoboken, NJ. All 1 rights reserved. In this lecture, you will learn: • Basic concepts of MPEG compression • Types of MPEG • Applications of MPEG • What GOP (group of pictures) in MPEG is • How GOP settings affect MPEG video file size • How GOP settings affect MPEG picture quality • What streaming video and progressive download are © 2016 Pearson Education, Inc., Hoboken, NJ. All 2 rights reserved. MPEG • Moving Pictures Experts Group Committee who derives standards for encoding video • Allow high compression • MPEG-1, MPEG-2, MPEG-4 © 2016 Pearson Education, Inc., Hoboken, NJ. All 3 rights reserved. What Happend to MPEG-3? • NOT MP3 (which is audio format) • Intended for HDTV • HDTV specifications was merged into MPEG-2 © 2016 Pearson Education, Inc., Hoboken, NJ. All 4 rights reserved. MPEG-1 • Video quality comparable to VHS • Originally intended for Web and CD-ROM playback • Frame sizes up to 352 × 240 pixels • Video format for VCD (VideoCD) before DVD became widespread © 2016 Pearson Education, Inc., Hoboken, NJ. All 5 rights reserved. MPEG-2 • Supports DVD-video, HDTV, HDV standards • For DVD video production: Export video into DVD MPEG-2 format • For HDV video production: Export video into HDV's MPEG-2 format © 2016 Pearson Education, Inc., Hoboken, NJ. All 6 rights reserved. MPEG-4 • Newer standard of MPEG family • Different encoding approach from MPEG-1 and MPEG-2 (will discuss after MPEG-1 and MPEG-2 compression in this lecture) © 2016 Pearson Education, Inc., Hoboken, NJ. All 7 rights reserved.
    [Show full text]
  • Application Note Introduction Video Basics
    Application Note Contents Title IP Video Encoding Explained Introduction...........................................................1 Series Understanding IP Video Performance Video Basics .........................................................1 Date May 2008 Video Codecs ........................................................2 Video Frame Types and the Group of Pictures............3 Overview Video Compression .................................................3 IP networks are increasingly used to deliver Encapsulation and Transmission ..........................5 video services for entertainment, education, business and commercial applications. Impairment Types .................................................6 Performance Management of IP Video.................8 This Application Note explains how video is encoded and transmitted over IP, describes Summary .............................................................10 common impairments that can affect video perceptual quality, and lists some basic approaches to IP video performance management. Introduction This Application Note will explain the video encoding process, as well as describing some common types of IP video (a.k.a. “Video over IP”) refers to any number impairments and performance management strategies of services in which a video signal is digitized, for ensuring the quality of IP-based video. compressed, and transmitted over an IP network to its destination, where it can then be decompressed, Video Basics decoded, and played. Typical services include IPTV, Video on Demand
    [Show full text]
  • About MPEG Compression HD Video Requires Significantly More Data Than SD Video
    About MPEG Compression HD video requires significantly more data than SD video. A single HD video frame can require up to six times more data than an SD frame. To record such large images with such a low data rate, HDV uses long-GOP MPEG compression. MPEG compression reduces the data rate by removing redundant visual information, both on a per-frame basis and also across multiple frames. Note: HDV specifically employs MPEG-2 compression, but the concepts of long-GOP and I-frame-only compression discussed below apply to all versions of the MPEG standard: MPEG-1, MPEG-2, and MPEG-4 (including AVC/H.264). For the purposes of this general explanation, the term MPEG may refer to any of these formats. Spatial Compression Within a single frame, areas of similar color and texture can be coded with fewer bits than the original frame, thus reducing the data rate with a minimal loss in noticeable visual quality. JPEG compression works in a similar way to compress still images. Spatial, or intraframe, compression is used to create standalone video frames called I-frames (short for intraframe). Temporal Compression Instead of storing complete frames, temporal (interframe) compression stores only what has changed from one frame to the next, which dramatically reduces the amount of data that needs to be stored while still achieving high-quality images. Video is stored in three types of frames: a standalone I-frame that contains a complete image, and then predictive P-frames and bipredictive B-frames that store subsequent changes in the image. Every half second or so, a new I-frame is introduced to provide a complete image on which subsequent P- and B-frames are based.
    [Show full text]