Efficient Rate Control Strategies for Scalable Video Coding

Total Page:16

File Type:pdf, Size:1020Kb

Efficient Rate Control Strategies for Scalable Video Coding Efficient rate control strategies for scalable video coding Thibaud Biatek To cite this version: Thibaud Biatek. Efficient rate control strategies for scalable video coding. Signal and Image process- ing. INSA de Rennes, 2016. English. NNT : 2016ISAR0007. tel-01392735 HAL Id: tel-01392735 https://tel.archives-ouvertes.fr/tel-01392735 Submitted on 4 Nov 2016 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. THESE INSA Rennes présentée par sous le sceau de l’Université Bretagne Loire pour obtenir le titre de Thibaud Biatek DOCTEUR DE L’INSA RENNES ECOLE DOCTORALE : MATISSE Spécialité : Traitement du Signal et des Images LABORATOIRE : IETR Thèse soutenue le 25.04.2016 Efficient Rate Control devant le jury composé de : Strategies for Scalable Christine Guillemot Directrice de Recherche à l’INRIA de Rennes (France) / Présidente Video Coding Marco Cagnazzo Maitre de Conférences HDR à Telecom-ParisTech (France) / Rapporteur Mathias Wien Chercheur à l’Université de Aachen (Allemagne) / Rapporteur Frédéric Dufaux Directeur de Recherche au CNRS LTCI, Paris (France) / Examinateur Patrick Le Callet Professeur à Polytech Nantes (France) / Examinateur Wassim Hamidouche Maitre de Conférences à l’INSA de Rennes (France) / Co-encadrant Jean-François Travers Expert Technique Systèmes Audiovisuels à TDF (France) / Co-encadrant Olivier Déforges Professeur à l’INSA de Rennes (France) / Directeur de thèse Stratégies d’Encodage pour Codeur Vidéo Scalable - Efficient Rate Control Strategies for Scalable Video Coding Thibaud Biatek En partenariat avec Document protégé par les droits d’auteur Acknowledgements First of all, I would like to thank one of my PhD supervisors Wassim Hamidouche. He has always been deeply involved, curious and critical about my work. He has always challenged me and helped me to be me more accurrate, meticulous, and ambitious with my research. I feel lucky to have been supervised by Wassim and I am convinced that I could not have achieved all this work without his valuable advice. I would also like to thank Jean-François Travers, my PhD supervisor at TDF. He was extensively helpful in bringing out the industrial relevance of my ideas and encouraged me to experiment my algorithms on pertinent use-cases, compliant with realistic broadcasting requirements. Our pleasant and interesting discussions at TDF were a rewarding experience, and I hope that our paths will cross again in the future. I also feel grateful to Mickael Raulet, who was my first contact in Rennes, and one of my supervisors during the first part of the thesis. His valuable advice helped me to publish and patent my first research results. Last but not least, this work is indebted to Olivier Déforges, my PhD surpervisor at the IETR. He has always been present to support and helped me in academic paper presentations. When I contemplate the progress made, I measure how my supervisors helped to be a better and more rigourous scientist today. I also feel grateful for all my TDF co-workers, especially Imad Alawe, Frederic Beauchamp, Philippe Courboulay, Pape-Abdoulaye Fam and Amelie Kerbellec for their support. I also would like to thank my jury: Marco Cagnazzo, Mathias Wien, Frédéric Dufaux, Christine Guillemot and Patrick Le Callet for accepting to evaluate my work. Their construc- tive feedback helped me improve the manuscript but also to point out new perspectives. During over two years, I worked within the Institute of Research and Technology b<>com, I do wish to thank all my co-workers, particularly Cambodge Bist, Pierre Castel, Franck Chi, Nicolas Dhollande, Antonin Gilles, Victorien Lorcy, Charlène Mouton, David Pesce, Adrien Schadle and Fabrice Templon for all the time spent and for the discussions we had together wether it was in or outside work. I would also like to thank Jean-Yves Aubie, Daniele Cloatre and Xavier Ducloux for their interest in my research. All these people definitely made Bcom a really inspiring and stimulating place towork. Finally, I would like to thank my parents Laurent and Marie-Pierre, my siblings Aurélien and Juliette and my fiancée Félicie for their invaluable support during these three years. Résumé en français Introduction High Efficiency Video Coding, ou HEVC/H.265, est la dernière norme de compression vidéo dont la première version a été publiée en avril 2013. Cette nouvelle norme est le fruit d’un travail collaboratif mené par des experts de l’ITU et de l’ISO/IEC respectivement représenté par les groupes VCEG (Video Coding Experts Group) et MPEG (Moving Pictures Experts Group), formant ainsi JCT-VC (Joint Collaborative Team on Video Coding). L’objectif de HEVC est de remplacer le très populaire Advanced Video Coding (AVC/H.264) dont l’utilisation est très largement répandue, en permettant de réduire son débit de codage de 50% pour une qualité visuelle équivalente. Comme son prédécesseur AVC l’a permis pour l’introduction de la Haute-Définition (HD), le déploiement d’HEVC permettra l’introduction de nouveaux services plus couteux en débit tel que l’Ultra-Haute-Définition (UHD). Actuellement, de nombreux réseaux servent de support à la transmission de contenus vidéo. Les diffuseurs historiques de la télévision numérique terrestre (TNT) perdent leurs fréquences au profit des opérateurs de téléphonie mobile. En France cette réduction du spectre hertzien s’est opéré via deux dividendes numérique au début des années 2000 puis en 2015. En plus du mobile et de la TNT, le réseau internet est devenu un support majeur de diffusion de contenus via les services over-the-top (OTT), la vidéo devrait représenter 80% du trafic en 2019. En plus de ces aspects réseaux, les supports de consommation des contenus vidéo se multiplient drastiquement. La télévision, les smartphones, tablettes, casques de réalités virtuelles et montre connectées sont les nouveaux supports et sont caractérisés par une multitude de formats différents (résolution, gamut de couleurs, dynamique, etc ...). Cette diversité implique une lourde redondance des flux encodés à de nombreux débits, résolution, qualités, et donc un coût supplémentaire selon plusieurs aspects. Premièrement pour la diffusion OTT, un même contenu est encodé de très nombreuses fois en de très nombreuses versions et à différents débits ce qui accroit les besoins de stockage. Du point de viii vue de la TNT, l’introduction de nouveaux formats multimédia est envisageable seulement en assurant une compatibilité avec le parc d’équipements existant. Cette compatibilité est très souvent assurée en opérant une diffusion dite simulcast, où toutes les versions d’un contenu sont transmises dans le canal, le récepteur décodant seulement la version qui le concerne. La compression vidéo scalable est une façon d’éviter tous ces inconvénients, et permet d’encoder un contenu en une couche de base et plusieurs couches d’améliorations de formats différents. Dans ce schéma, la compression est significativement plus efficace que le SimulCast puisque chaque couche est encodée en exploitant des informations provenant des couches inférieures. La norme Scalable High Efficiency Video Coding, ou SHVC, a été introduite dans les Annexes F et H d’HEVC et publié en octobre 2014 dans la version 2 de HEVC. SHVC cou- vre plusieurs types de scalabilité, spatiale, en qualité, en gamut de couleur et en codec en plus du mécanisme de scalabilité temporelle déjà inclus dans la première version d’HEVC. Ces types de scalabilité peuvent être combinés entre eux et peuvent résoudre les problématiques de diffusion actuelles. Parmis ces problèmes, l’introduction de nouveaux services UHD rétrocompatibles avec le parc existant via les scalabilité HD/UHD, BT.709 vers BT.2020, SDR vers HDR. Cette technologie est déjà considérée dans plusieurs groupes de normali- sation tel que l’Advanced Television Systems Committee (ATSC) et le consortium Digital Video Broadcasting (DVB). Dans la norme ATSC3.0, SHVC est la technologie retenue pour la diffusion de programme scalable (limitée à deux couches spatiales). Du côté DVB la scalabilité temporelle est déjà retenue dans la norme ETSI TS-101-154 pour la diffusion de programme UHD, et SHVC est considéré comme candidat pour la diffusion rétro-compatible des futurs formats en cours de normalisation. Pour la diffusion de services scalables dans un environnement industriel, le contrôle de débit sur les flux générés est un enjeu majeur. Pour cette raison, l’objectif de cette thèseaété le développement d’algorithmes de régulation de débit pour SHVC. Dans un premier volet, nous allons explorer des algorithmes d’estimation du paramètre de quantification (QP) basée sur l’approche r-domaine afin d’atteindre un débit cible lors de l’encodage. Deuxièmement, nous développerons des algorithmes d’allocation de débit adaptatifs pour SHVC qui aurons pour but d’ajuster le ratio de débit entre les couches de façon à optimiser les performances de compression. Finalement, nous investiguerons le multiplexage statistique de programmes SHVC encodés dans le contexte d’introduction de services UHD. ix Etat de l’art Dans le Chapitre2, les principes de compression vidéo ainsi qu’HEVC, ses outils et son extension scalable SHVC sont introduits ainsi qu’un aperçu des logiciels de référence utilisés en normalisation. HEVC, comme AVC avant lui, est basé sur un schéma de codage hybride par bloc. Les images en entrée du codeur sont découpées en blocs appelés Coding Tree Unit (CTU) et employant une représentation en arbre de sous-blocs Coding Unit (CU). Chacun de ces CU est partitionné en blocs sur lesquels la prédiction est réalisée (Prediction Units ou PU) et d’autres adaptés aux étapes de transformation et quantification (Transform Units ou TU).
Recommended publications
  • Kulkarni Uta 2502M 11649.Pdf
    IMPLEMENTATION OF A FAST INTER-PREDICTION MODE DECISION IN H.264/AVC VIDEO ENCODER by AMRUTA KIRAN KULKARNI Presented to the Faculty of the Graduate School of The University of Texas at Arlington in Partial Fulfillment of the Requirements for the Degree of MASTER OF SCIENCE IN ELECTRICAL ENGINEERING THE UNIVERSITY OF TEXAS AT ARLINGTON May 2012 ACKNOWLEDGEMENTS First and foremost, I would like to take this opportunity to offer my gratitude to my supervisor, Dr. K.R. Rao, who invested his precious time in me and has been a constant support throughout my thesis with his patience and profound knowledge. His motivation and enthusiasm helped me in all the time of research and writing of this thesis. His advising and mentoring have helped me complete my thesis. Besides my advisor, I would like to thank the rest of my thesis committee. I am also very grateful to Dr. Dongil Han for his continuous technical advice and financial support. I would like to acknowledge my research group partner, Santosh Kumar Muniyappa, for all the valuable discussions that we had together. It helped me in building confidence and motivated towards completing the thesis. Also, I thank all other lab mates and friends who helped me get through two years of graduate school. Finally, my sincere gratitude and love go to my family. They have been my role model and have always showed me right way. Last but not the least; I would like to thank my husband Amey Mahajan for his emotional and moral support. April 20, 2012 ii ABSTRACT IMPLEMENTATION OF A FAST INTER-PREDICTION MODE DECISION IN H.264/AVC VIDEO ENCODER Amruta Kulkarni, M.S The University of Texas at Arlington, 2011 Supervising Professor: K.R.
    [Show full text]
  • Download Media Player Codec Pack Version 4.1 Media Player Codec Pack
    download media player codec pack version 4.1 Media Player Codec Pack. Description: In Microsoft Windows 10 it is not possible to set all file associations using an installer. Microsoft chose to block changes of file associations with the introduction of their Zune players. Third party codecs are also blocked in some instances, preventing some files from playing in the Zune players. A simple workaround for this problem is to switch playback of video and music files to Windows Media Player manually. In start menu click on the "Settings". In the "Windows Settings" window click on "System". On the "System" pane click on "Default apps". On the "Choose default applications" pane click on "Films & TV" under "Video Player". On the "Choose an application" pop up menu click on "Windows Media Player" to set Windows Media Player as the default player for video files. Footnote: The same method can be used to apply file associations for music, by simply clicking on "Groove Music" under "Media Player" instead of changing Video Player in step 4. Media Player Codec Pack Plus. Codec's Explained: A codec is a piece of software on either a device or computer capable of encoding and/or decoding video and/or audio data from files, streams and broadcasts. The word Codec is a portmanteau of ' co mpressor- dec ompressor' Compression types that you will be able to play include: x264 | x265 | h.265 | HEVC | 10bit x265 | 10bit x264 | AVCHD | AVC DivX | XviD | MP4 | MPEG4 | MPEG2 and many more. File types you will be able to play include: .bdmv | .evo | .hevc | .mkv | .avi | .flv | .webm | .mp4 | .m4v | .m4a | .ts | .ogm .ac3 | .dts | .alac | .flac | .ape | .aac | .ogg | .ofr | .mpc | .3gp and many more.
    [Show full text]
  • (A/V Codecs) REDCODE RAW (.R3D) ARRIRAW
    What is a Codec? Codec is a portmanteau of either "Compressor-Decompressor" or "Coder-Decoder," which describes a device or program capable of performing transformations on a data stream or signal. Codecs encode a stream or signal for transmission, storage or encryption and decode it for viewing or editing. Codecs are often used in videoconferencing and streaming media solutions. A video codec converts analog video signals from a video camera into digital signals for transmission. It then converts the digital signals back to analog for display. An audio codec converts analog audio signals from a microphone into digital signals for transmission. It then converts the digital signals back to analog for playing. The raw encoded form of audio and video data is often called essence, to distinguish it from the metadata information that together make up the information content of the stream and any "wrapper" data that is then added to aid access to or improve the robustness of the stream. Most codecs are lossy, in order to get a reasonably small file size. There are lossless codecs as well, but for most purposes the almost imperceptible increase in quality is not worth the considerable increase in data size. The main exception is if the data will undergo more processing in the future, in which case the repeated lossy encoding would damage the eventual quality too much. Many multimedia data streams need to contain both audio and video data, and often some form of metadata that permits synchronization of the audio and video. Each of these three streams may be handled by different programs, processes, or hardware; but for the multimedia data stream to be useful in stored or transmitted form, they must be encapsulated together in a container format.
    [Show full text]
  • Opus, a Free, High-Quality Speech and Audio Codec
    Opus, a free, high-quality speech and audio codec Jean-Marc Valin, Koen Vos, Timothy B. Terriberry, Gregory Maxwell 29 January 2014 Xiph.Org & Mozilla What is Opus? ● New highly-flexible speech and audio codec – Works for most audio applications ● Completely free – Royalty-free licensing – Open-source implementation ● IETF RFC 6716 (Sep. 2012) Xiph.Org & Mozilla Why a New Audio Codec? http://xkcd.com/927/ http://imgs.xkcd.com/comics/standards.png Xiph.Org & Mozilla Why Should You Care? ● Best-in-class performance within a wide range of bitrates and applications ● Adaptability to varying network conditions ● Will be deployed as part of WebRTC ● No licensing costs ● No incompatible flavours Xiph.Org & Mozilla History ● Jan. 2007: SILK project started at Skype ● Nov. 2007: CELT project started ● Mar. 2009: Skype asks IETF to create a WG ● Feb. 2010: WG created ● Jul. 2010: First prototype of SILK+CELT codec ● Dec 2011: Opus surpasses Vorbis and AAC ● Sep. 2012: Opus becomes RFC 6716 ● Dec. 2013: Version 1.1 of libopus released Xiph.Org & Mozilla Applications and Standards (2010) Application Codec VoIP with PSTN AMR-NB Wideband VoIP/videoconference AMR-WB High-quality videoconference G.719 Low-bitrate music streaming HE-AAC High-quality music streaming AAC-LC Low-delay broadcast AAC-ELD Network music performance Xiph.Org & Mozilla Applications and Standards (2013) Application Codec VoIP with PSTN Opus Wideband VoIP/videoconference Opus High-quality videoconference Opus Low-bitrate music streaming Opus High-quality music streaming Opus Low-delay
    [Show full text]
  • Spb Tv Encoder
    SPB TV ENCODER INPUT IP Input Interfaces 2x Gigabit Ethernet SPB TV Encoder is a professional solution for the fast, high- that processes multi-format video streams for delivery to different quality transcoding of live and on-demand video streams from end-user devices: Mobile, Tablet, Desktop, and connected TV-sets. Optional Digital Input Interfaces SDI, DVB-S a single headend to multi-screen devices. Leverage your network SPB TV Encoding solution can be delivered in a flexible software- Input IP Streams UDP Unicast/Multicast, HTTP, HLS, RTSP, RTMP, MMS with a unified platform for IP TV, OTT TV and mobile TV services based configuration or as off the-shelf equipment. Supported Codecs MPEG-2, MPEG-4, H.264, MPEG-1, MPEG-1 Layer 3, AAC, HE AAC, AC3, WMV 7, WMV 8, WMV 9, WMA 1, WMA 2, WMA 9 Pro, On2 VP3, On2 VP5, On2 VP6, On2 VP8, PNG Filters Video cropping, Image insertion on input source loss, automatic audio loudness adjustment Up to 100 multi-rate channels FEC (Forward Error Correction) OUTPUT Live and on-demand video encoding 5.1 Audio, auto audio gain support IP Output Interfaces 2x Gigabit Ethernet Remote monitoring tools Keeping aspect ratio Output IP streams RTP Unicast/Multicast Image insertion on input loss Image overlay Output Codecs H.264 (Baseline, Main, High), MPEG-4, AAC, HE AAC, H.265 Web GUI and CLI interfaces Hardsubs overlay VoD Output Multi Track MP4 VoD Publishing FTP, SFTP SNMP 3D encoding support Thumbnails PNG, JPEG Managed updates to include the latest Multi output support codecs, formats, profiles and devices MANAGEMENT/MONITORING
    [Show full text]
  • Voxpro/Audionlabs Was Acquired by Wheatstone Corporation in the Fall of 2015
    NOTE: VoxPro/AudionLabs was acquired by Wheatstone Corporation in the fall of 2015. As a result the contact information and support links listed in this legacy document are no longer current. Should you require further assistance regarding the information presented here, please refer to the following contact info: EMAIL: [email protected] TEL: +1.252-638-7000 (9am-5pm EST) WEB: wheatstone.com [menu item: VoxPro] MAIL: Wheatstone Corporation 600 Industrial Drive New Bern, NC. 28562 USA 100716 Contents 3 Table of Contents 0 Chapter 1 Introduction 8 Chapter 2 Getting Started 10 1 VoxPro Main................................................................................................................................... Window 11 2 VoxPro Control................................................................................................................................... Panel 12 3 User Accounts................................................................................................................................... 13 4 Passwords ................................................................................................................................... 14 Chapter 3 Recording 16 1 Create New ...................................................................................................................................Recording 16 2 Create Empty................................................................................................................................... Recording 18 3 Create Recording..................................................................................................................................
    [Show full text]
  • Multiple Reference Motion Compensation: a Tutorial Introduction and Survey Contents
    Foundations and TrendsR in Signal Processing Vol. 2, No. 4 (2008) 247–364 c 2009 A. Leontaris, P. C. Cosman and A. M. Tourapis DOI: 10.1561/2000000019 Multiple Reference Motion Compensation: A Tutorial Introduction and Survey By Athanasios Leontaris, Pamela C. Cosman and Alexis M. Tourapis Contents 1 Introduction 248 1.1 Motion-Compensated Prediction 249 1.2 Outline 254 2 Background, Mosaic, and Library Coding 256 2.1 Background Updating and Replenishment 257 2.2 Mosaics Generated Through Global Motion Models 261 2.3 Composite Memories 264 3 Multiple Reference Frame Motion Compensation 268 3.1 A Brief Historical Perspective 268 3.2 Advantages of Multiple Reference Frames 270 3.3 Multiple Reference Frame Prediction 271 3.4 Multiple Reference Frames in Standards 277 3.5 Interpolation for Motion Compensated Prediction 281 3.6 Weighted Prediction and Multiple References 284 3.7 Scalable and Multiple-View Coding 286 4 Multihypothesis Motion-Compensated Prediction 290 4.1 Bi-Directional Prediction and Generalized Bi-Prediction 291 4.2 Overlapped Block Motion Compensation 294 4.3 Hypothesis Selection Optimization 296 4.4 Multihypothesis Prediction in the Frequency Domain 298 4.5 Theoretical Insight 298 5 Fast Multiple-Frame Motion Estimation Algorithms 301 5.1 Multiresolution and Hierarchical Search 302 5.2 Fast Search using Mathematical Inequalities 303 5.3 Motion Information Re-Use and Motion Composition 304 5.4 Simplex and Constrained Minimization 306 5.5 Zonal and Center-biased Algorithms 307 5.6 Fractional-pixel Texture Shifts or Aliasing
    [Show full text]
  • The Daala Video Codec Project Next-Next Generation Video
    The Daala Video Codec Project Next-next Generation Video Timothy B. Terriberry Mozilla & The Xiph.Org Foundation ● Patents are no longer a problem for free software – We can all go home 2 Mozilla & The Xiph.Org Foundation ● Except... not quite 3 Mozilla & The Xiph.Org Foundation Carving out Exceptions in OIN (Table 0 contains one Xiph codec: FLAC) 4 Mozilla & The Xiph.Org Foundation Why This Matters ● Encumbered codecs are a billion dollar toll-tax on communications – Every cost from codecs is repeated a million fold in all multimedia software ● Codec licensing is anti-competitive – Licensing regimes are universally discriminatory – An excuse for proprietary software (Flash) ● Ignoring licensing creates risks that can show up at any time – A tax on success 5 Mozilla & The Xiph.Org Foundation The Royalty-Free Video Challenge ● Creating good codecs is hard – But we don’t need many – The best implementations of patented codecs are already free software ● Network effects decide – Where RF is established, non-free codecs see no adoption (JPEG, PNG, FLAC, …) ● RF is not enough – People care about different things – Must be better on all fronts 6 Mozilla & The Xiph.Org Foundation We Did This for Audio 7 Mozilla & The Xiph.Org Foundation The Daala Project ● Goal: Better than HEVC without infringing IPR ● Need a better strategy than “read a lot of patents” – People don’t believe you – Analysis is error-prone ● Try to stay far away from the line, but... ● One mistake can ruin years of development effort ● See: H.264 Baseline 8 Mozilla & The Xiph.Org
    [Show full text]
  • PVQ Applied Outside of Daala (IETF 97 Draft)
    PVQ Applied outside of Daala (IETF 97 Draft) Yushin Cho Mozilla Corporation November, 2016 Introduction ● Perceptual Vector Quantization (PVQ) – Proposed as a quantization and coefficient coding tool for NETVC – Originally developed for the Daala video codec – Does a gain-shape coding of transform coefficients ● The most distinguishing idea of PVQ is the way it references a predictor. – PVQ does not subtract the predictor from the input to produce a residual Mozilla Corporation 2 Integrating PVQ into AV1 ● Introduction of a transformed predictors both in encoder and decoder – Because PVQ references the predictor in the transform domain, instead of using a pixel-domain residual as in traditional scalar quantization ● Activity masking, the major benefit of PVQ, is not enabled yet – Encoding RDO is solely based on PSNR Mozilla Corporation 3 Traditional Architecture Input X residue Subtraction Transform T signal R predictor P + decoded X Inverse Inverse Scalar decoded R Transform Quantizer Quantizer Coefficient bitstream of Coder coded T(X) Mozilla Corporation 4 AV1 with PVQ Input X Transform T T(X) PVQ Quantizer PVQ Coefficient predictor P Transform T T(X) Coder PVQ Inverse Quantizer Inverse dequantized bitstream of decoded X Transform T(X) coded T(X) Mozilla Corporation 5 Coding Gain Change Metric AV1 --> AV1 with PVQ PSNR Y -0.17 PSNR-HVS 0.27 SSIM 0.93 MS-SSIM 0.14 CIEDE2000 -0.28 ● For the IETF test sequence set, "objective-1-fast". ● IETF and AOM for high latency encoding options are used. Mozilla Corporation 6 Speed ● Increase in total encoding time due to PVQ's search for best codepoint – PVQ's time complexity is close to O(n*n) for n coefficients, while scalar quantization has O(n) ● Compared to Daala, the search space for a RDO decision in AV1 is far larger – For the 1st frame of grandma_qcif (176x144) in intra frame mode, Daala calls PVQ 3843 times, while AV1 calls 632,520 times, that is ~165x.
    [Show full text]
  • CALIFORNIA STATE UNIVERSITY, NORTHRIDGE Optimized AV1 Inter
    CALIFORNIA STATE UNIVERSITY, NORTHRIDGE Optimized AV1 Inter Prediction using Binary classification techniques A graduate project submitted in partial fulfillment of the requirements for the degree of Master of Science in Software Engineering by Alex Kit Romero May 2020 The graduate project of Alex Kit Romero is approved: ____________________________________ ____________ Dr. Katya Mkrtchyan Date ____________________________________ ____________ Dr. Kyle Dewey Date ____________________________________ ____________ Dr. John J. Noga, Chair Date California State University, Northridge ii Dedication This project is dedicated to all of the Computer Science professors that I have come in contact with other the years who have inspired and encouraged me to pursue a career in computer science. The words and wisdom of these professors are what pushed me to try harder and accomplish more than I ever thought possible. I would like to give a big thanks to the open source community and my fellow cohort of computer science co-workers for always being there with answers to my numerous questions and inquiries. Without their guidance and expertise, I could not have been successful. Lastly, I would like to thank my friends and family who have supported and uplifted me throughout the years. Thank you for believing in me and always telling me to never give up. iii Table of Contents Signature Page ................................................................................................................................ ii Dedication .....................................................................................................................................
    [Show full text]
  • A/300, "ATSC 3.0 System Standard"
    ATSC A/300:2017 ATSC 3.0 System 19 October 2017 ATSC Standard: ATSC 3.0 System (A/300) Doc. A/300:2017 19 October 2017 Advanced Television Systems Committee 1776 K Street, N.W. Washington, DC 20006 202-872-9160 i ATSC A/300:2017 ATSC 3.0 System 19 October 2017 The Advanced Television Systems Committee, Inc., is an international, non-profit organization developing voluntary standards for digital television. The ATSC member organizations represent the broadcast, broadcast equipment, motion picture, consumer electronics, computer, cable, satellite, and semiconductor industries. Specifically, ATSC is working to coordinate television standards among different communications media focusing on digital television, interactive systems, and broadband multimedia communications. ATSC is also developing digital television implementation strategies and presenting educational seminars on the ATSC standards. ATSC was formed in 1982 by the member organizations of the Joint Committee on InterSociety Coordination (JCIC): the Electronic Industries Association (EIA), the Institute of Electrical and Electronic Engineers (IEEE), the National Association of Broadcasters (NAB), the National Cable Telecommunications Association (NCTA), and the Society of Motion Picture and Television Engineers (SMPTE). Currently, there are approximately 120 members representing the broadcast, broadcast equipment, motion picture, consumer electronics, computer, cable, satellite, and semiconductor industries. ATSC Digital TV Standards include digital high definition television (HDTV), standard definition television (SDTV), data broadcasting, multichannel surround-sound audio, and satellite direct-to-home broadcasting. Note: The user’s attention is called to the possibility that compliance with this standard may require use of an invention covered by patent rights. By publication of this standard, no position is taken with respect to the validity of this claim or of any patent rights in connection therewith.
    [Show full text]
  • Video Compression Optimized for Racing Drones
    Video compression optimized for racing drones Henrik Theolin Computer Science and Engineering, master's level 2018 Luleå University of Technology Department of Computer Science, Electrical and Space Engineering Video compression optimized for racing drones November 10, 2018 Preface To my wife and son always! Without you I'd never try to become smarter. Thanks to my supervisor Staffan Johansson at Neava for providing room, tools and the guidance needed to perform this thesis. To my examiner Rickard Nilsson for helping me focus on the task and reminding me of the time-limit to complete the report. i of ii Video compression optimized for racing drones November 10, 2018 Abstract This thesis is a report on the findings of different video coding tech- niques and their suitability for a low powered lightweight system mounted on a racing drone. Low latency, high consistency and a robust video stream is of the utmost importance. The literature consists of multiple comparisons and reports on the efficiency for the most commonly used video compression algorithms. These reports and findings are mainly not used on a low latency system but are testing in a laboratory environment with settings unusable for a real-time system. The literature that deals with low latency video streaming and network instability shows that only a limited set of each compression algorithms are available to ensure low complexity and no added delay to the coding process. The findings re- sulted in that AVC/H.264 was the most suited compression algorithm and more precise the x264 implementation was the most optimized to be able to perform well on the low powered system.
    [Show full text]