Fast VP9-To-AV1 Transcoding Based on Block Partitioning Inheritance

Total Page:16

File Type:pdf, Size:1020Kb

Fast VP9-To-AV1 Transcoding Based on Block Partitioning Inheritance Fast VP9-to-AV1 Transcoding based on Block Partitioning Inheritance Alex Borges, Daniel Palomino, Bruno Zatt, Marcelo Porto, Guilherme Correa Video Technology Research Group (ViTech) Graduate Program in Computing (PPGC), Federal University of Pelotas (UFPel), Brazil {amborges, dpalomino, zatt, porto, gcorrea}@inf.ufpel.edu.br Abstract— This paper proposes a fast VP9-to-AV1 video compared to VP9 (considering the same image quality). This transcoding algorithm based on block partitioning inheritance. represents an average superiority of 22% for AV1 over VP9, The proposed algorithm relies on the reuse of VP9 block an economy of 1/5 in storage resources and other costs partitioning during the AV1 re-encoding process. This way, the involving video transmission services. exhaustive search for the best block size option is avoided to save encoding time. The reuse of VP9 block partitioning is proposed Video transcoding is the process that converts from one based on a statiscal analysis that shows the relation of block video bitstream format to the same format with different parititioning sizes between VP9 and AV1. The analysis configurations (homogeneous transcoding) or to another demontrates that there is a high probability of the AV1 encoding bitstream format (heterogeneous transcoding), as presented in process to choose block sizes of the same size as in the VP9 Fig. 1. With the advent of AV1, converting legacy content encoding. Experimental results show that the proposed from previous formats, such as VP9, becomes an essential task algorithm is able to accelerate the VP9-to-AV1 transcoding for service providers that intend to benefit from its process by 28% on average at the cost of only 4% increase in the compression efficiency. However, as the computational cost BD-Rate when compared with the complete decoding and re- required by libaom is too high, speeding up the encoding encoding process. process is important to allow fast transcoding without decreasing compression efficiency significantly. Keywords—AV1, VP9, transcoding, video coding VP9 and AV1 share several characteristics that can be I. INTRODUCTION harnessed during the transcoding process. Both VP9 and AV1 In 2015, the Alliance for Open Media (AOMedia) follow a block-based hybrid video coding process, as they consortium was created to develop modern royalty-free video divide the frame into smaller parts called blocks for coding formats for online applications, such as on-demand prediction, transform, and quantization. Blocks can assume video transmission, videoconferences, and live streaming. different sizes and shapes, as defined by the codec Partially based on the VP9 [1], Daala [2] and Thor [3] codecs, specification. To achieve the best compression efficiency, the AOMedia launched the AOMedia Video 1 (AV1) [4] format. encoder needs to find the best block size to use in each region Along with the specification, the libaom [5] reference of the video. Usually, this is performed by exhaustive search software was released in 2018. Since then, many other fast over all block size possibilities, which requires a huge AV1 codecs versions have been developed and released by computational effort. In [13], a method for inheriting the best AOMedia members, such as the Scalable Video Technology block size from H.264/AVC to H.265/HEVC during the for AV1 Encoder (SVT-AV1) [6] developed by Intel and transcoding process is proposed. In [14], the H.265/HEVC Netflix, the CISCO-AV1 [7] developed by CISCO, and the Coding Unit depth information is used to accelerate the rav1e [8] developed by XIPH. transcoding from H.265/HEVC to AV1. Although the methods proposed in [13] and [14] present good results and One of the main goals of AV1 is to overcome the are based on block partitioning inheritance, they focus on the compression efficiency achieved by VP9 and replace it as H.264/AVC and H.265/HEVC standards, which employ a current state-of-the-art technology based on royalty-free very different set of block sizes, partitioning modes and codecs. To accomplish that, AV1 includes several new tools coding tree structure in comparison to VP9. Thus, they cannot and features with much more efficient signal processing be directly employed to accelerate the VP9-to-AV1 operations and frame partitioning structures in comparison to transcoding process. To the best of the authors’ knowledge, VP9. However, this efficiency is achieved at the cost of a there is no other work focusing on VP9-to-AV1 transcoding. considerable complexity increase in comparison to VP9. The authors in [9] and [10] show that the reference libaom codec This paper proposes a fast VP9-to-AV1 transcoding requires an encoding time more than 100 times larger in process based on Block Partitioning Inheritance. The comparison to VP9. Thus, time-saving strategies for AV1, proposed solution saves time by reusing the VP9 block especially those leading to small or no penalties in terms of partitioning direction to filter out AV1 partitioning compression efficiency, are currently essential to reduce this possibilities during the re-encoding process. The idea is based gap and enable the deployment of AV1 codecs. on a statistical analysis performed over a set of VP9 and AV1 bitstreams, which allowed identifying partition modes rarely VP9 owner and developer, Google, is the main company used under certain circumstances. that makes use of VP9 in its video services, like the YouTube platform [11], one of the most popular free video platforms in the world. According to [12] more than 500 hours of video content around the world is published Video Transcoder Bitstream Bitstream every minute on YouTube. All these videos are stored in *.vp9 VP9 AV1 decoder decoded coder *.av1 large data centers and require a huge space in hard drives. VIDEO In [9], [10], and [4] the authors demonstrate that AV1 can achieve a compression efficiency gain of 20%, 18%, Fig. 1. Tandem VP9-to-AV1 transcoder. and 28%, respectively, when 978-9-0827-9705-3 555 EUSIPCO 2020 A. Experimental Setup The Spatial Information and Temporal Information (SI- TI) analysis [15] was performed over all the test sequences available in [16], section “objective-2-slow”, to identify those with most heterogeneous characteristics in order to enable a diverse set of videos to be used for testing. Videos sequences selected for the statistical analysis were Blue Sky, BQ Highway, Dirt Bike, Guitar HDR Amazon, Netflix Dinner Scene, Netflix Food Market 2, Netflix Tunnel Flag, and Water HDR Amazon, as available in [17]. To perform the Fig. 2. Block partitioning allowed in AV1 and VP9 (highlighted). experiments, 60 frames of each video sequence were encoded. The reference codec software for both VP9 and AV1 was used in the experiments. For VP9, the libvpx [18] codec, 128x128 version 1.8.2 (hash code 50d1a4), was used. For AV1, the libaom [5] codec, version 1.0.0 (hash code db8f27), was used. 64x64 The reference software implementations were chosen because 32x32 they represent the most complete versions of the encoder 16x16 specifications, including all the available modes and partitioning possibilities allowed in both formats. 8x8 IETF-NETVC-T [16] is the documentation that defines the 4x4 test configurations used for both video codecs. Following the document, the High Latency CQP configuration was used in (a) (b) the experiments and the Constrained Quality (CQ) parameter was set to values 20, 32, 43, and 55. All experiments are Fig. 3. An AV1 superblock partitioned into blocks: (a) block view, (b) tree view. Gray blocks represent Split partitioning mode. executed sequentially in the same workstation (Intel [email protected] GHz processor, 8 GB RAM, Ubuntu OS), in terminal mode. II. VP9/AV1 PARTITIONING CORRELATION ANALYSIS The CQ parameter was set to 20 for quantization in the A block can assume square or rectangular shapes VP9 encoder, aiming at transcoding from the best quality organized in some configurations, called partitions type. In available in the recommended settings. For AV1, CQ values VP9 there are three partition types: square (named as None), 20, 32, 43, and 55 were used, as defined in [16]. As AV1 vertical (Vert) and horizontal (Horz), as shown in Fig. 2. In introduces new partitioning possibilities in both the horizontal AV1, blocks can assume nine partition types based on the and vertical directions, the occurrence rate of modes three directions observed in VP9. Besides, both codecs also belonging to each direction was summed up in the analysis. allow a Split partition type, which recursively subdivides the Thus, modes 2, 3, 4, and 8 in Fig. 2 are considered as Horz current block into four square blocks. This process follows a modes, whereas modes 1, 5, 6, and 7 are considered as Vert coding block tree structure, as shown in Fig. 3. It is likely that modes. the same block partitioning will be used in both VP9 and AV1 B. Correlation Analysis codecs, since the same video region is being encoded. Considering this, we performed a correlation analysis between For each block in a VP9-encoded video, the same region the partitioning chosen by each codec to use it as a basis for in the AV1-encoded video was observed. For that, a label was the proposed fast transcoding algorithm. attributed to each 4×4 area in the frame, indicating which block size and partition mode were chosen during the TABLE I. CORRELATION ANALYSIS BETWEEN PARTITION TYPES CHOSEN BY VP9 AND AV1 (CQ 20). VP9 64×64 32×32 16×16 8×8 AV1 None Horz Vert None Horz Vert None Horz Vert None Horz Vert None 39.18 25.16 10.92 4.56 1.86 1.72 1.61
Recommended publications
  • Network Working Group A
    Network Working Group A. Filippov Internet Draft Huawei Technologies Intended status: Informational A. Norkin Netflix J.R. Alvarez Huawei Technologies Expires: May 17, 2017 November 17, 2016 <Video Codec Requirements and Evaluation Methodology> draft-ietf-netvc-requirements-04.txt Status of this Memo This Internet-Draft is submitted in full conformance with the provisions of BCP 78 and BCP 79. Internet-Drafts are working documents of the Internet Engineering Task Force (IETF), its areas, and its working groups. Note that other groups may also distribute working documents as Internet- Drafts. The list of current Internet-Drafts can be accessed at http://datatracker.ietf.org/drafts/current/ Internet-Drafts are draft documents valid for a maximum of six months and may be updated, replaced, or obsoleted by other documents at any time. It is inappropriate to use Internet-Drafts as reference material or to cite them other than as "work in progress." The list of current Internet-Drafts can be accessed at http://www.ietf.org/1id-abstracts.html The list of Internet-Draft Shadow Directories can be accessed at http://www.ietf.org/shadow.html This Internet-Draft will expire on May 17, 2017. Copyright Notice Copyright (c) 2016 IETF Trust and the persons identified as the document authors. All rights reserved. This document is subject to BCP 78 and the IETF Trust’s Legal Provisions Relating to IETF Documents (http://trustee.ietf.org/license-info) in effect on the date of publication of this document. Please review these documents carefully, as they describe your rights and restrictions with <Filippov> Expires May 17, 2017 [Page 1] Internet-Draft Video Codec Requirements and Evaluation November 2016 respect to this document.
    [Show full text]
  • Thumbor-Video-Engine Release 1.2.2
    thumbor-video-engine Release 1.2.2 Aug 14, 2021 Contents 1 Installation 3 2 Setup 5 2.1 Contents.................................................6 2.2 License.................................................. 13 2.3 Indices and tables............................................ 13 i ii thumbor-video-engine, Release 1.2.2 thumbor-video-engine provides a thumbor engine that can read, crop, and transcode audio-less video files. It supports input and output of animated GIF, animated WebP, WebM (VP9) video, and MP4 (default H.264, but HEVC is also supported). Contents 1 thumbor-video-engine, Release 1.2.2 2 Contents CHAPTER 1 Installation pip install thumbor-video-engine Go to GitHub if you need to download or install from source, or to report any issues. 3 thumbor-video-engine, Release 1.2.2 4 Chapter 1. Installation CHAPTER 2 Setup In your thumbor configuration file, change the ENGINE setting to 'thumbor_video_engine.engines. video' to enable video support. This will allow thumbor to support video files in addition to whatever image formats it already supports. If the file passed to thumbor is an image, it will use the Engine specified by the configuration setting IMAGING_ENGINE (which defaults to 'thumbor.engines.pil'). To enable transcoding between formats, add 'thumbor_video_engine.filters.format' to your FILTERS setting. If 'thumbor.filters.format' is already present, replace it with the filter from this pack- age. ENGINE = 'thumbor_video_engine.engines.video' FILTERS = [ 'thumbor_video_engine.filters.format', 'thumbor_video_engine.filters.still', ] To enable automatic transcoding to animated gifs to webp, you can set FFMPEG_GIF_AUTO_WEBP to True. To use this feature you cannot set USE_GIFSICLE_ENGINE to True; this causes thumbor to bypass the custom ENGINE altogether.
    [Show full text]
  • The Interplay of Compile-Time and Run-Time Options for Performance Prediction Luc Lesoil, Mathieu Acher, Xhevahire Tërnava, Arnaud Blouin, Jean-Marc Jézéquel
    The Interplay of Compile-time and Run-time Options for Performance Prediction Luc Lesoil, Mathieu Acher, Xhevahire Tërnava, Arnaud Blouin, Jean-Marc Jézéquel To cite this version: Luc Lesoil, Mathieu Acher, Xhevahire Tërnava, Arnaud Blouin, Jean-Marc Jézéquel. The Interplay of Compile-time and Run-time Options for Performance Prediction. SPLC 2021 - 25th ACM Inter- national Systems and Software Product Line Conference - Volume A, Sep 2021, Leicester, United Kingdom. pp.1-12, 10.1145/3461001.3471149. hal-03286127 HAL Id: hal-03286127 https://hal.archives-ouvertes.fr/hal-03286127 Submitted on 15 Jul 2021 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. The Interplay of Compile-time and Run-time Options for Performance Prediction Luc Lesoil, Mathieu Acher, Xhevahire Tërnava, Arnaud Blouin, Jean-Marc Jézéquel Univ Rennes, INSA Rennes, CNRS, Inria, IRISA Rennes, France [email protected] ABSTRACT Both compile-time and run-time options can be configured to reach Many software projects are configurable through compile-time op- specific functional and performance goals. tions (e.g., using ./configure) and also through run-time options (e.g., Existing studies consider either compile-time or run-time op- command-line parameters, fed to the software at execution time).
    [Show full text]
  • Screen Capture Tools to Record Online Tutorials This Document Is Made to Explain How to Use Ffmpeg and Quicktime to Record Mini Tutorials on Your Own Computer
    Screen capture tools to record online tutorials This document is made to explain how to use ffmpeg and QuickTime to record mini tutorials on your own computer. FFmpeg is a cross-platform tool available for Windows, Linux and Mac. Installation and use process depends on your operating system. This info is taken from (Bellard 2016). Quicktime Player is natively installed on most of Mac computers. This tutorial focuses on Linux and Mac. Table of content 1. Introduction.......................................................................................................................................1 2. Linux.................................................................................................................................................1 2.1. FFmpeg......................................................................................................................................1 2.1.1. installation for Linux..........................................................................................................1 2.1.1.1. Add necessary components........................................................................................1 2.1.2. Screen recording with FFmpeg..........................................................................................2 2.1.2.1. List devices to know which one to record..................................................................2 2.1.2.2. Record screen and audio from your computer...........................................................3 2.2. Kazam........................................................................................................................................4
    [Show full text]
  • DVP Tutorial
    IP Video Conferencing: A Tutorial Roman Sorokin, Jean-Louis Rougier Abstract Video conferencing is a well-established area of communications, which have been studied for decades. Recently this area has received a new impulse due to significantly increased bandwidth of Local and Wide area networks, appearance of low-priced video equipment and development of web based media technologies. This paper presents the main techniques behind the modern IP-based videoconferencing services, with a particular focus on codecs, network protocols, architectures and standardization efforts. Questions of security and topologies are also tackled. A description of a typical video conference scenario is provided, demonstrating how the technologies, responsible for different conference aspects, are working together. Traditional industrial disposition as well as modern innovative approaches are both addressed. Current industry trends are highlighted in respect to the topics, described in the tutorial. Legacy analog/digital technologies, together with the gateways between the traditional and the IP videoconferencing systems, are not considered. Keywords Video Conferencing, codec, SVC, MCU, SIP, RTP Roman Sorokin ALE International, Colombes, France e-mail: [email protected] Jean-Louis Rougier Télécom ParisTech, Paris, France e-mail: [email protected] 1 1 Introduction Video conferencing is a two-way interactive communication, delivered over networks of different nature, which allows people from several locations to participate in a meeting. Conference participants use video conferencing endpoints of different types. Generally a video conference endpoint has a camera and a microphone. The video stream, generated by the camera, and the audio stream, coming from the microphone, are both compressed and sent to the network interface.
    [Show full text]
  • (A/V Codecs) REDCODE RAW (.R3D) ARRIRAW
    What is a Codec? Codec is a portmanteau of either "Compressor-Decompressor" or "Coder-Decoder," which describes a device or program capable of performing transformations on a data stream or signal. Codecs encode a stream or signal for transmission, storage or encryption and decode it for viewing or editing. Codecs are often used in videoconferencing and streaming media solutions. A video codec converts analog video signals from a video camera into digital signals for transmission. It then converts the digital signals back to analog for display. An audio codec converts analog audio signals from a microphone into digital signals for transmission. It then converts the digital signals back to analog for playing. The raw encoded form of audio and video data is often called essence, to distinguish it from the metadata information that together make up the information content of the stream and any "wrapper" data that is then added to aid access to or improve the robustness of the stream. Most codecs are lossy, in order to get a reasonably small file size. There are lossless codecs as well, but for most purposes the almost imperceptible increase in quality is not worth the considerable increase in data size. The main exception is if the data will undergo more processing in the future, in which case the repeated lossy encoding would damage the eventual quality too much. Many multimedia data streams need to contain both audio and video data, and often some form of metadata that permits synchronization of the audio and video. Each of these three streams may be handled by different programs, processes, or hardware; but for the multimedia data stream to be useful in stored or transmitted form, they must be encapsulated together in a container format.
    [Show full text]
  • Opus, a Free, High-Quality Speech and Audio Codec
    Opus, a free, high-quality speech and audio codec Jean-Marc Valin, Koen Vos, Timothy B. Terriberry, Gregory Maxwell 29 January 2014 Xiph.Org & Mozilla What is Opus? ● New highly-flexible speech and audio codec – Works for most audio applications ● Completely free – Royalty-free licensing – Open-source implementation ● IETF RFC 6716 (Sep. 2012) Xiph.Org & Mozilla Why a New Audio Codec? http://xkcd.com/927/ http://imgs.xkcd.com/comics/standards.png Xiph.Org & Mozilla Why Should You Care? ● Best-in-class performance within a wide range of bitrates and applications ● Adaptability to varying network conditions ● Will be deployed as part of WebRTC ● No licensing costs ● No incompatible flavours Xiph.Org & Mozilla History ● Jan. 2007: SILK project started at Skype ● Nov. 2007: CELT project started ● Mar. 2009: Skype asks IETF to create a WG ● Feb. 2010: WG created ● Jul. 2010: First prototype of SILK+CELT codec ● Dec 2011: Opus surpasses Vorbis and AAC ● Sep. 2012: Opus becomes RFC 6716 ● Dec. 2013: Version 1.1 of libopus released Xiph.Org & Mozilla Applications and Standards (2010) Application Codec VoIP with PSTN AMR-NB Wideband VoIP/videoconference AMR-WB High-quality videoconference G.719 Low-bitrate music streaming HE-AAC High-quality music streaming AAC-LC Low-delay broadcast AAC-ELD Network music performance Xiph.Org & Mozilla Applications and Standards (2013) Application Codec VoIP with PSTN Opus Wideband VoIP/videoconference Opus High-quality videoconference Opus Low-bitrate music streaming Opus High-quality music streaming Opus Low-delay
    [Show full text]
  • High Efficiency, Moderate Complexity Video Codec Using Only RF IPR
    Thor High Efficiency, Moderate Complexity Video Codec using only RF IPR draft-fuldseth-netvc-thor-00 Arild Fuldseth, Gisle Bjontegaard (Cisco) IETF 93 – Prague, CZ – July 2015 1 Design principles • Moderate complexity to allow real-time implementation in SW on common HW, as well as new HW designs • Basic building blocks from well-known hybrid approach (motion compensated prediction and transform coding) • Common design elements in modern codecs – Larger block sizes and transforms, up to 64x64 – Quarter pixel interpolation, motion vector prediction, etc. • Cisco RF IPR (note well: declaration filed on draft) – Deblocking, transforms, etc. (some also essential in H.265/4) • Avoid non-RF IPR – If/when others offer RF IPR, design/performance will improve 2 Encoder Architecture Input Transform Quantizer Entropy Output video Coding bitstream - Inverse Transform Intra Frame Prediction Loop filters Inter Frame Prediction Reconstructed Motion Frame Estimation Memory 3 Decoder Architecture Input Entropy Inverse Bitstream Decoding Transform Intra Frame Prediction Loop filters Inter Frame Prediction Output video Reconstructed Frame Memory 4 Block Structure • Super block (SB) 64x64 • Quad-tree split into coding blocks (CB) >= 8x8 • Multiple prediction blocks (PB) per CB • Intra: 1 PB per CB • Inter: 1, 2 (rectangular) or 4 (square) PBs per CB • 1 or 4 transform blocks (TB) per CB 5 Coding-block modes • Intra • Inter0 MV index, no residual information • Inter1 MV index, residual information • Inter2 Explicit motion vector information, residual information
    [Show full text]
  • Encoding H.264 Video for Streaming and Progressive Download
    W4: KEY ENCODING SKILLS, TECHNOLOGIES TECHNIQUES STREAMING MEDIA EAST - 2019 Jan Ozer www.streaminglearningcenter.com [email protected]/ 276-235-8542 @janozer Agenda • Introduction • Lesson 5: How to build encoding • Lesson 1: Delivering to Computers, ladder with objective quality metrics Mobile, OTT, and Smart TVs • Lesson 6: Current status of CMAF • Lesson 2: Codec review • Lesson 7: Delivering with dynamic • Lesson 3: Delivering HEVC over and static packaging HLS • Lesson 4: Per-title encoding Lesson 1: Delivering to Computers, Mobile, OTT, and Smart TVs • Computers • Mobile • OTT • Smart TVs Choosing an ABR Format for Computers • Can be DASH or HLS • Factors • Off-the-shelf player vendor (JW Player, Bitmovin, THEOPlayer, etc.) • Encoding/transcoding vendor Choosing an ABR Format for iOS • Native support (playback in the browser) • HTTP Live Streaming • Playback via an app • Any, including DASH, Smooth, HDS or RTMP Dynamic Streaming iOS Media Support Native App Codecs H.264 (High, Level 4.2), HEVC Any (Main10, Level 5 high) ABR formats HLS Any DRM FairPlay Any Captions CEA-608/708, WebVTT, IMSC1 Any HDR HDR10, DolbyVision ? http://bit.ly/hls_spec_2017 iOS Encoding Ladders H.264 HEVC http://bit.ly/hls_spec_2017 HEVC Hardware Support - iOS 3 % bit.ly/mobile_HEVC http://bit.ly/glob_med_2019 Android: Codec and ABR Format Support Codecs ABR VP8 (2.3+) • Multiple codecs and ABR H.264 (3+) HLS (3+) technologies • Serious cautions about HLS • DASH now close to 97% • HEVC VP9 (4.4+) DASH 4.4+ Via MSE • Main Profile Level 3 – mobile HEVC (5+)
    [Show full text]
  • Arxiv:2007.15943V1 [Cs.SE] 31 Jul 2020
    MUZZ: Thread-aware Grey-box Fuzzing for Effective Bug Hunting in Multithreaded Programs Hongxu Chen§† Shengjian Guo‡ Yinxing Xue§∗ Yulei Sui¶ Cen Zhang† Yuekang Li† Haijun Wang# Yang Liu† †Nanyang Technological University ‡Baidu Security ¶University of Technology Sydney §University of Science and Technology of China #Ant Financial Services Group Abstract software performance. A typical computing paradigm of mul- tithreaded programs is to accept a set of inputs, distribute Grey-box fuzz testing has revealed thousands of vulner- computing jobs to threads, and orchestrate their progress ac- abilities in real-world software owing to its lightweight cordingly. Compared to sequential programs, however, multi- instrumentation, fast coverage feedback, and dynamic adjust- threaded programs are more prone to severe software faults. ing strategies. However, directly applying grey-box fuzzing On the one hand, the non-deterministic thread-interleavings to input-dependent multithreaded programs can be extremely give rise to concurrency-bugs like data-races, deadlocks, inefficient. In practice, multithreading-relevant bugs are usu- etc [32]. These bugs may cause the program to end up with ab- ally buried in the sophisticated program flows. Meanwhile, normal results or unexpected hangs. On the other hand, bugs existing grey-box fuzzing techniques do not stress thread- that appear under specific inputs and interleavings may lead interleavings that affect execution states in multithreaded pro- to concurrency-vulnerabilities [5, 30], resulting in memory grams. Therefore, mainstream grey-box fuzzers cannot ade- corruptions, information leakage, etc. quately test problematic segments in multithreaded software, although they might obtain high code coverage statistics. There exist a line of works on detecting bugs and vulner- To this end, we propose MUZZ, a new grey-box fuzzing abilities in multithreaded programs.
    [Show full text]
  • Hardware for Speech and Audio Coding
    Linköping Studies in Science and Technology Thesis No. 1093 Hardware for Speech and Audio Coding Mikael Olausson LiU-TEK-LIC-2004:22 Department of Electrical Engineering Linköpings universitet, SE-581 83 Linköping, Sweden Linköping 2004 Linköping Studies in Science and Technology Thesis No. 1093 Hardware for Speech and Audio Coding Mikael Olausson LiU-TEK-LIC-2004:22 Department of Electrical Engineering Linköpings universitet, SE-581 83 Linköping, Sweden Linköping 2004 ISBN 91-7373-953-7 ISSN 0280-7971 ii Abstract While the Micro Processors (MPUs) as a general purpose CPU are converging (into Intel Pentium), the DSP processors are diverging. In 1995, approximately 50% of the DSP processors on the market were general purpose processors, but last year only 15% were general purpose DSP processors on the market. The reason general purpose DSP processors fall short to the application specific DSP processors is that most users want to achieve highest performance under mini- mized power consumption and minimized silicon costs. Therefore, a DSP proces- sor must be an Application Specific Instruction set Processor (ASIP) for a group of domain specific applications. An essential feature of the ASIP is its functional acceleration on instruction level, which gives the specific instruction set architecture for a group of appli- cations. Hardware acceleration for digital signal processing in DSP processors is essential to enhance the performance while keeping enough flexibility. In the last 20 years, researchers and DSP semiconductor companies have been working on different kinds of accelerations for digital signal processing. The trade-off be- tween the performance and the flexibility is always an interesting question because all DSP algorithms are "application specific"; the acceleration for audio may not be suitable for the acceleration of baseband signal processing.
    [Show full text]
  • IP-Soc Shanghai 2017 ALLEGRO Presentation FINAL
    Building an Area-optimized Multi-format Video Encoder IP Tomi Jalonen VP Sales www.allegrodvt.com Allegro DVT Founded in 2003 Privately owned, based in Grenoble (France) Two product lines: 1) Industry de-facto standard video compliance streams Decoder syntax, performance and error resilience streams for H.264|MVC, H.265/SHVC, VP9, AVS2 and AV1 System compliance streams 2) Leading semiconductor video IP Multi-format encoder IP for H.264, H.265, VP9, JPEG Multi-format decoder IP for H.264, H.265, VP9, JPEG WiGig IEEE 802.11ad WDE CODEC IP 2 Evolution of Video Coding Standards International standards defined by standardization bodies such as ITU-T and ISO/IEC H.261 (1990) MPEG-1 (1993) H.262 / MPEG-2 (1995) H.263 (1996) MPEG-4 Part 2 (1999) H.264 / AVC / MPEG-4 Part 10 (2003) H.265 / HEVC (2013) Future Video Coding (“FVC”) MPEG and ISO "Preliminary Joint Call for Evidence on Video Compression with Capability beyond HEVC.” (202?) Incremental improvements of transform-based & motion- compensated hybrid video coding schemes to meet the ever increasing resolution and frame rate requirements 3 Regional Video Standards SMPTE standards in the US VC-1 (2006) VC-2 (2008) China Information Industry Department standards AVS (2005) AVS+ (2012) AVS2.0 (2016) 4 Proprietary Video Formats Sorenson Spark On2 VP6, VP7 RealVideo DivX Popular in the past partly due to technical merits but mainly due to more suitable licensing schemes to a given application than standard video video formats with their patent royalties. 5 Royalty-free Video Formats Xiph.org Foundation
    [Show full text]