Low-Complexity Adaptive Block-Size Transform Based on Extended Transforms

Total Page:16

File Type:pdf, Size:1020Kb

Low-Complexity Adaptive Block-Size Transform Based on Extended Transforms LOW-COMPLEXITY ADAPTIVE BLOCK-SIZE TRANSFORM BASED ON EXTENDED TRANSFORMS Honggang Qi1,Wen Gao 1, Siwei Ma2, Debin Zhao 1 and Xiangyang Ji1 1Institute of Computing Technology, Chinese Academy of Sciences E-mail: {hgqi, wgao, dbzhao, xyji}@jdl.ac.cn 2Department of Electrical Engineering and Integrated Medias Center University of Southern California E-mail: swma@usc.edu ABSTRACT small transform prevents more ringing artifacts in coded pictures [2]. To exert the virtues of different transform In this paper, a low-complexity 8x8/4x4 Adaptive Block- sizes greatly, Adaptive Block-size Transform (ABT) size Transform (ABT) scheme is tentatively proposed for scheme is employed in image and video coding standards. the Chinese Audio and Video coding Standard (AVS)1. The scheme can improve the coding performance In the proposed ABT scheme, an integer 8x8 transform is significantly. For example, the average luma PSNR gain derived from the integer 4x4 transform used in AVS in high-definition sequences is up to 0.5dB in H.264/AVC. according to a transform extension principle. The 8x8 In ABT scheme, the transforms in different sizes are transform not only has high energy compacted property adaptively chosen by encoder. Usually, RD decision is but also can be merely implemented within several employed in encoder for choosing the better transform [2]. additions and shifts, and all intermediate results are With RD decision, the optimal coding performance can be limited with 16-bit. The 8x8 transform and 4x4 transform achieved in RD sense. Since two transform sizes are can be merged together and share the same scale matrix so employed in ABT scheme, two zig-zag scans are needed that the hardware units and storage resources are for corresponding transforms. Moreover, the RD decision efficiently saved for both encoder and decoder. The requires that the current block is coded twice with the two experimental results on numerous sequences show that the transforms. It is clear that the ABT scheme introduces proposed ABT scheme can achieves significant high complexity to encoder. However, in view of its high performance improvement for AVS. coding performance, the increased complexity is tolerable in some applications. Thus, ABT scheme is adopted into 1. INTRODUCTION Fidelity Range Extensions of H.264/AVC by Joint Video Team (JVT). The Discrete Cosine Transform (DCT) is an effective In this paper, the ABT scheme with extended transform coding tool for image and video coding. In transforms is proposed for Chinese Audio and Video hybrid video coding, it is used to convert the correlated coding Standard (AVS). In this scheme, the extended 8x8 prediction residuals to seldom correlated frequency transform is derived from the original 4x4 transform in domain coefficients for more efficient data compression. AVS. The extended 8x8 transform not only improves the Different coding standards employ different transforms. coding performance of ABT but also reduces the memory MPGE-2 adopts an 8x8 floating-point DCT, which is a requirements and hardware units for both encoder and real DCT with fine compression performance but high decoder. Since the computation and storage complexities complexity. H.264/AVC adopts a 4x4 integer DCT [1], of ABT mainly focus on the encoder and only some which has similar properties to real 4x4 DCT but relative storage complexities are increased in the decoder, the low complexity. Since the inverse DCT in the encoder and proposed method can reduce the complexities of decoder the decoder are defined by exact integer operations, the efficiently. Thus, the method is meaningful to some problem of mismatches is avoided. applications where the decoders are paid more attention The different transform sizes possess different than encoder. coding performances. Larger transform has a better The rest of this paper is organized as follows: In the energy compaction than smaller transform. However, the section 2, the concept of extended transforms is introduced, and an 8x8 transform which is an extension of 1 4x4 transform of AVS is designed. In section 3, the The Chinese audio and video coding standard, an emerging standard targeting at higher coding efficiency and lower complexity. implementation of ABT in AVS is briefly specified. In 1­4244­0367­7/06/$20.00 ©2006 IEEE 129 ICME 2006 section 4, the experimental results on test sequences with A low-complexity integer 4x4 transform is employed in different coding characteristics are shown. Finally, the AVS-M (a version of AVS orient to mobile applications), conclusions are made by us. The transform is expressed as follows: 2. THE EXTENDED TRANSFORMS OF ABT ª 2 2 2 2º « 3 1 1 3» Extended transforms means that the NxN transform as a C4 « » (3) part involves in the 2Nx2N transform [4]. Taking 4x4 and « 2 2 2 2» « » 8x8 transform as an example, if the 4x4 transform is ¬ 1 3 3 1¼ expressed with (1), its extended 8x8 transform should be expressed as (2). The transform has a good energy compaction and low complexity. It can be implemented with merely several ª c0 c0 c0 c0 º addition and shift operations. In practice, the C4 is just a « » c2 c6 c6 c2 (1) core matrix of DCT. The norms of its some rows are D4 « » « c4 c4 c4 c4 » unequal. A scale matrix is needed for normalizing the core « » matrix. As the output of the transform and scale of the ¬ c6 c2 c2 c6 ¼ input X, Y is expressed as ª c c c c c c c c º 0 0 0 0 0 0 0 0 Y (C4u X uC4c) S4 (4) « c c c c c c c c » « 1 3 5 7 7 5 3 1» « c2 c6 c6 c2 c2 c6 c6 c2 » « » where the notation denotes element-by-element c c c c c c c c (2) D8 « 3 7 1 5 5 1 7 3 » multiplication, and S4 means the 4x4 scale matrix. The « c c c c c c c c » « 4 4 4 4 4 4 4 4 » scale matrix S4 can be calculated by « c5 c1 c7 c3 c3 c7 c1 c5 » « c c c c c c c c » 6 2 2 6 6 2 2 6 3 3 « » 2 2 ¬« c7 c5 c3 c1 c1 c3 c5 c7 ¼» S4x,y 1 ( ¦ C4x,j )u( ¦C4ic,y ) (5) j 0 i 0 It can be found from (1) and (2) that the even basis functions of 8x8 transform D8 inherit all basis functions 2.2. 8x8 transform design of 4x4 transform D4. i.e. D82i,j= D4i,j, (0İi, j<4). The relation can be more clearly expressed from their butterfly According to the requirement of performance, the integer flow graph in Fig.1, where the two 4x4 transforms D4II 8x8 transform should preserve the real DCT compaction and D4IV mean the II-type and IV-type DCT respectively property. Its basis functions of transform matrix should [3]. If D4II is identical to D4, it is said that D8 is extended approximate the real transform’s. Meanwhile, these basis from D4. functions should not be so large that the transform can be implemented within narrow bit range. Moreover, the integer 8x8 transform should be an extension of the integer 4x4 transform C4 so that the 8x8 transform can be shared by the 4x4/8x8 transforms. Thus, the integer 8x8 II D4x4 transform should meet the following three design rules: 1) The basis functions of each row of the designed 8x8 transform should correlate closely with corresponding floating-point basis functions of real 8x8 transform where the correlation is measured by the cosine of the included angle between them; D IV 4x4 2) The basis functions of the designed 8x8 transform matrix should be represented by several simple integers; 3) The designed 8x8 transform must be the extension of Fig.1. Fast DCT butterfly signal flow graph. 4x4 transform of AVS i.e. meeting the relation: C82x,y=C4x,y, 0İx, y<4. 2.1. 4x4 transform 130 According to the above rules, an integer 8x8 DCT is prediction mode. Moreover, numorous experimental expressed as follows: results show that ABT applying to chroma signal does not bring the significant coding gains. Thus, from the view ª 4 4 4 4 4 4 4 4º point of complexityˈ ABT scheme is not used in chroma « 6 6 3 2 2 3 6 6» signals. Unlike the conventional schemes where all 8x8, « » « 6 2 2 6 6 2 2 6» 8x4, 4x8 and 4x4 transforms are used, no more transforms « » are used in proposed scheme except for 8x8 and 4x4 6 2 6 3 3 6 2 6 C8 « » / 2 (6) transforms. Because the 8x4 and 4x8 transforms just « 4 4 4 4 4 4 4 4» « » obtain small PSNR gains while introducing large coding « 3 6 2 6 6 2 6 3» loads to the codec. « 2 6 6 2 2 6 6 2» RD optimizer is used to choose the transform size « » with the better coding performance for current coded « 2 3 6 6 6 6 3 2» ¬ ¼ block. Since the RD optimizer just works in the encoder, a one-bit flag is added into bitstream to signal the decoder where C8 is an integer orthogonal matrix. Its even basis which transform is used to coding the current block. For functions and those of integer 4x4 transform matrix C4 further reducing the complexity, the 4x4 and 8x8 zig-zag meet the relation, C82x,y=C4x,y. Thus, C8 is an extended scans, entropy codings and deblocking tools are also transform of C4. The odd basis functions of C4 consists of merged with some optimization algorithms. For example, integers (6, 6, 3, 2), which are so simple that dynamic the 8x8 quantized transform coefficients are split into four range is only increased by 5.1 bits (log234).
Recommended publications
  • Bosch Intelligent Streaming
    Intelligent streaming | Bitrate-optimized high-quality video 1 | 22 Intelligent streaming Bitrate-optimized high-quality video – White Paper Data subject to change without notice | May 31st, 2019 BT-SC Intelligent streaming | Bitrate-optimized high-quality video 2 | 22 Table of contents 1 Introduction 3 2 Basic structures and processes 4 2.1 Image capture process ..................................................................................................................................... 4 2.2 Image content .................................................................................................................................................... 4 2.3 Scene content .................................................................................................................................................... 5 2.4 Human vision ..................................................................................................................................................... 5 2.5 Video compression ........................................................................................................................................... 6 3 Main concepts and components 7 3.1 Image processing .............................................................................................................................................. 7 3.2 VCA ..................................................................................................................................................................... 7 3.3 Smart Encoder
    [Show full text]
  • Isize Bitsave:High-Quality Video Streaming at Lower Bitrates
    Datasheet v20210518 iSIZE BitSave: High-Quality Video Streaming at Lower Bitrates iSIZE’s innovative BitSave technology comprises an AI-based perceptual preprocessing solution that allows conventional, third-party encoders to produce higher quality video at lower bitrate. BitSave achieves this by preprocessing the video content before it reaches the video encoder (Fig. 1), such that the output after compressing with any standard video codec is perceptually optimized with less motion artifacts or blurring, for the same or lower bitrate. BitSave neural networks are able to isolate areas of perceptual importance, such as those with high motion or detailed texture, and optimize their structure so that they are preserved better at lower bitrate by any subsequent encoder. As shown in our recent paper, when integrated as a preprocessor prior to a video encoder, BitSave is able to reduce bitrate requirements for a given quality level by up to 20% versus that encoder. When coupled with open-source AVC, HEVC and AV1 encoders, BitSave allows for up to 40% bitrate reduction at no quality compromise in comparison to leading third-party AVC, HEVC and AV1 encoding services. Lower bitrates equate to smaller filesizes, lower storage, transmission and distribution costs, reduced energy consumption and more satisfied customers. Source BitSave Encoder Delivery AI-based pre- Single pass processing processing prior One frame latency per content for an to encoding (AVC, entire ABR ladder HEVC, VP9, AV1) Improves encoding quality Integrated within Intel as measured by standard OpenVINO, ONNX and Dolby perceptual quality metrics Vision, easy to plug&play (VMAF, SSIM, VIF) within any existing workflow Figure 1.
    [Show full text]
  • Towards Optimal Compression of Meteorological Data: a Case Study of Using Interval-Motivated Overestimators in Global Optimization
    Towards Optimal Compression of Meteorological Data: A Case Study of Using Interval-Motivated Overestimators in Global Optimization Olga Kosheleva Department of Electrical and Computer Engineering and Department of Teacher Education, University of Texas, El Paso, TX 79968, USA, olgak@utep.edu 1 Introduction The existing image and data compression techniques try to minimize the mean square deviation between the original data f(x; y; z) and the compressed- decompressed data fe(x; y; z). In many practical situations, reconstruction that only guaranteed mean square error over the data set is unacceptable. For example, if we use the meteorological data to plan a best trajectory for a plane, then what we really want to know are the meteorological parameters such as wind, temperature, and pressure along the trajectory. If along this line, the values are not reconstructed accurately enough, the plane may crash; the fact that on average, we get a good reconstruction, does not help. In general, what we need is a compression that guarantees that for each (x; y), the di®erence jf(x; y; z)¡fe(x; y; z)j is bounded by a given value ¢, i.e., that the actual value f(x; y; z) belongs to the interval [fe(x; y; z) ¡ ¢; fe(x; y; z) + ¢]: In this chapter, we describe new e±cient techniques for data compression under such interval uncertainty. 2 Optimal Data Compression: A Problem 2.1 Data Compression Is Important At present, so much data is coming from measuring instruments that it is necessary to compress this data before storing and processing.
    [Show full text]
  • Adaptive Bitrate Streaming Over Cellular Networks: Rate Adaptation and Data Savings Strategies
    Adaptive Bitrate Streaming Over Cellular Networks: Rate Adaptation and Data Savings Strategies Yanyuan Qin, Ph.D. University of Connecticut, 2021 ABSTRACT Adaptive bitrate streaming (ABR) has become the de facto technique for video streaming over the Internet. Despite a flurry of techniques, achieving high quality ABR streaming over cellular networks remains a tremendous challenge. First, the design of an ABR scheme needs to balance conflicting Quality of Experience (QoE) metrics such as video quality, quality changes, stalls and startup performance, which is even harder under highly dynamic bandwidth in cellular network. Second, streaming providers have been moving towards using Variable Bitrate (VBR) encodings for the video content, which introduces new challenges for ABR streaming, whose nature and implications are little understood. Third, mobile video streaming consumes a lot of data. Although many video and network providers currently offer data saving options, the existing practices are suboptimal in QoE and resource usage. Last, when the audio and video arXiv:2104.01104v2 [cs.NI] 14 May 2021 tracks are stored separately, video and audio rate adaptation needs to be dynamically coordinated to achieve good overall streaming experience, which presents interesting challenges while, somewhat surprisingly, has received little attention by the research community. In this dissertation, we tackle each of the above four challenges. Firstly, we design a framework called PIA (PID-control based ABR streaming) that strategically leverages PID control concepts and novel approaches to account for the various requirements of ABR streaming. The evaluation results demonstrate that PIA outperforms state-of-the-art schemes in providing high average bitrate with signif- icantly lower bitrate changes and stalls, while incurring very small runtime overhead.
    [Show full text]
  • Download Full Text (Pdf)
    Bitrate Requirements of Non-Panoramic VR Remote Rendering Viktor Kelkkanen Markus Fiedler David Lindero Blekinge Institute of Technology Blekinge Institute of Technology Ericsson Karlskrona, Sweden Karlshamn, Sweden Luleå, Sweden viktor.kelkkanen@bth.se markus.fiedler@bth.se david.lindero@ericsson.com ABSTRACT This paper shows the impact of bitrate settings on objective quality measures when streaming non-panoramic remote-rendered Virtual Reality (VR) images. Non-panoramic here refers to the images that are rendered and sent across the network, they only cover the viewport of each eye, respectively. To determine the required bitrate of remote rendering for VR, we use a server that renders a 3D-scene, encodes the resulting images using the NVENC H.264 codec and transmits them to the client across a network. The client decodes the images and displays them in the VR headset. Objective full-reference quality measures are taken by comparing the image before encoding on the server to the same image after it has been decoded on the client. By altering the average bitrate setting of the encoder, we obtain objective quality Figure 1: Example of the non-panoramic view and 3D scene scores as functions of bitrates. Furthermore, we study the impact that is rendered once per eye and streamed each frame. of headset rotation speeds, since this will also have a large effect on image quality. We determine an upper and lower bitrate limit based on headset 1 INTRODUCTION rotation speeds. The lower limit is based on a speed close to the Remote rendering for VR is commonly done by using panorama average human peak head-movement speed, 360°/s.
    [Show full text]
  • An Analysis of VP8, a New Video Codec for the Web Sean Cassidy
    Rochester Institute of Technology RIT Scholar Works Theses Thesis/Dissertation Collections 11-1-2011 An Analysis of VP8, a new video codec for the web Sean Cassidy Follow this and additional works at: http://scholarworks.rit.edu/theses Recommended Citation Cassidy, Sean, "An Analysis of VP8, a new video codec for the web" (2011). Thesis. Rochester Institute of Technology. Accessed from This Thesis is brought to you for free and open access by the Thesis/Dissertation Collections at RIT Scholar Works. It has been accepted for inclusion in Theses by an authorized administrator of RIT Scholar Works. For more information, please contact ritscholarworks@rit.edu. An Analysis of VP8, a New Video Codec for the Web by Sean A. Cassidy A Thesis Submitted in Partial Fulfillment of the Requirements for the Degree of Master of Science in Computer Engineering Supervised by Professor Dr. Andres Kwasinski Department of Computer Engineering Rochester Institute of Technology Rochester, NY November, 2011 Approved by: Dr. Andres Kwasinski R.I.T. Dept. of Computer Engineering Dr. Marcin Łukowiak R.I.T. Dept. of Computer Engineering Dr. Muhammad Shaaban R.I.T. Dept. of Computer Engineering Thesis Release Permission Form Rochester Institute of Technology Kate Gleason College of Engineering Title: An Analysis of VP8, a New Video Codec for the Web I, Sean A. Cassidy, hereby grant permission to the Wallace Memorial Library to repro- duce my thesis in whole or part. Sean A. Cassidy Date i Acknowledgements I would like to thank my thesis advisors, Dr. Andres Kwasinski, Dr. Marcin Łukowiak, and Dr. Muhammad Shaaban for their suggestions, advice, and support.
    [Show full text]
  • Encoding Parameters Prediction for Convex Hull Video Encoding
    Encoding Parameters Prediction for Convex Hull Video Encoding Ping-Hao Wu∗, Volodymyr Kondratenko y, Gaurang Chaudhari z, and Ioannis Katsavounidis z Facebook Inc., 1 Hacker Way, Menlo Park, CA 94025 Email: npw@fb.com∗, vkondratenko@fb.comy, gaurangc@fb.comz, ikatsavounidis@fb.comx Abstract—Fast encoding parameter selection technique have BD-rate loss. It uses a faster encoder or the same encoder been proposed in the past. Leveraging the power of convex hull with faster speed settings to perform analysis on the video video encoding framework, an encoder with a faster speed setting, sequence first, to determine the optimal encoding parameters or faster encoder such as hardware encoder, can be used to determine the optimal encoding parameters. It has been shown to be used with a slower encoder with better compression that one can speed up 3 to 5 times, while still achieve 20-40% efficiency. Although this fast selection technique can achieve BD-rate savings, compared to the traditional fixed-QP encodings. similar savings with a fraction of the computations, limitations Such approach presents two problems. First, although the are still present. First of all, the few percent loss in BD-rate is speedup is impressive, there is still ∼3% loss in BD-rate. not negligible in the case of higher-quality contents, where a Secondly, the previous approach only works with encoders implementing standards that have the same quantization scheme much higher target quality and/or target bitrate is often used. and QP range, such as between VP9 and AV1, and not in the Secondly, the quantization scheme of the fast encoder, used scenario where one might want to use a much faster H.264 for analysis, and that of the slow encoder, used for producing encoder, to determine and predict the encoding parameters for the actual deliverable encodings, need to be the same in order VP9 or AV1 encoder.
    [Show full text]
  • H.264 Codec Comparison
    MPEG-4 AVC/H.264 Video Codecs Comparison Short version of report Project head: Dr. Dmitriy Vatolin Measurements, analysis: Dmitriy Kulikov, Alexander Parshin Codecs: XviD (MPEG-4 ASP codec) MainConcept H.264 Intel H.264 x264 AMD H.264 Artemis H.264 December 2007 CS MSU Graphics&Media Lab Video Group http://www.compression.ru/video/codec_comparison/index_en.html videocodec-testing@graphics.cs.msu.ru VIDEO MPEG-4 AVC/H.264 CODECS COMPARISON MOSCOW, DEC 2007 CS MSU GRAPHICS & MEDIA LAB VIDEO GROUP SHORT VERSION Contents 1 Acknowledgments ................................................................................................4 2 Overview ..............................................................................................................5 2.1 Difference between Short and Full Versions............................................................ 5 2.2 Sequences ............................................................................................................... 5 2.3 Codecs ..................................................................................................................... 6 3 Objectives and Testing Rules ..............................................................................7 3.1 H.264 Codec Testing Objectives.............................................................................. 7 3.2 Testing Rules ........................................................................................................... 7 4 Comparison Results.............................................................................................9
    [Show full text]
  • Technologies for 3D Mesh Compression: a Survey
    J. Vis. Commun. Image R. 16 (2005) 688–733 www.elsevier.com/locate/jvci Technologies for 3D mesh compression: A survey Jingliang Peng a, Chang-Su Kim b,*, C.-C. Jay Kuo a a Integrated Media Systems Center, Department of Electrical Engineering, University of Southern California, Los Angeles, CA 90089-2564, USA b Department of Information Engineering, The Chinese University of Hong Kong, Shatin, Hong Kong Received 1 February 2004; accepted 5 March 2005 Available online 16 April 2005 Abstract Three-dimensional (3D) meshes have been widely used in graphic applications for the rep- resentation of 3D objects. They often require a huge amount of data for storage and/or trans- mission in the raw data format. Since most applications demand compact storage, fast transmission, and efficient processing of 3D meshes, many algorithms have been proposed to compress 3D meshes efficiently since early 1990s. In this survey paper, we examine 3D mesh compression technologies developed over the last decade, with the main focus on triangular mesh compression technologies. In this effort, we classify various algorithms into classes, describe main ideas behind each class, and compare the advantages and shortcomings of the algorithms in each class. Finally, we address some trends in the 3D mesh compression tech- nology development. Ó 2005 Elsevier Inc. All rights reserved. Keywords: 3D mesh compression; Single-rate mesh coding; Progressive mesh coding; MPEG-4 1. Introduction Graphics data are more and more widely used in various applications, including video gaming, engineering design, architectural walkthrough, virtual reality, e-com- * Corresponding author. E-mail addresses: jingliap@usc.edu (J.
    [Show full text]
  • Delivering Live and On-Demand Smooth Streaming
    Delivering Live and On-Demand Smooth Streaming Video delivered over the Internet has become an increasingly important element of the online presence of many companies. Organizations depend on Internet video for everything from corporate training and product launch webcasts, to more complex streamed content such as live televised sports events. No matter what the scale, to be successful an online video program requires a carefully-planned workflow that employs the right technology and tools, network architecture, and event management. This deployment guide was developed by capturing real world experiences from iStreamPlanet® and its use of the Microsoft® platform technologies, including Microsoft® SilverlightTM and Internet Information Services (IIS) Smooth Streaming, that have powered several successful online video events. This document aims to deliver an introductory view of the considerations for acquiring, creating, delivering, and managing live and on-demand SD and HD video. This guide will touch on a wide range of new tools and technologies which are emerging and essential components to deployed solutions, including applications and tools for live and on- demand encoding, delivery, content management (CMS) and Digital Asset Management (DAM) editing, data synchronization, and advertisement management. P1 | Delivering Live and On-Demand Streaming Media A Step-by-Step Development Process This paper was developed to be a hands-on, practical 1. Production document to provide guidance, best practices, and 2. Content Acquisition resources throughout every step of the video production 3. Encoding and delivery workflow. When delivering a live event over 4. Content Delivery the Internet, content owners need to address six key 5. Playback Experience areas, shown in Figure 1: 6.
    [Show full text]
  • Lossless Coding of Point Cloud Geometry Using a Deep Generative Model
    1 Lossless Coding of Point Cloud Geometry using a Deep Generative Model Dat Thanh Nguyen, Maurice Quach, Student Member, IEEE, Giuseppe Valenzise, Senior Member, IEEE, Pierre Duhamel, Life Fellow, IEEE Abstract—This paper proposes a lossless point cloud (PC) been optimized over several decades. On the other hand, G- geometry compression method that uses neural networks to PCC targets static content, and the geometry and attribute estimate the probability distribution of voxel occupancy. First, information are independently encoded. Color attributes can to take into account the PC sparsity, our method adaptively partitions a point cloud into multiple voxel block sizes. This be encoded using methods based on the Region Adaptive partitioning is signalled via an octree. Second, we employ a Hierarchical Transform (RAHT) [4], Predicting Transform or deep auto-regressive generative model to estimate the occupancy Lifting Transform [3]. Coding the PC geometry is particularly probability of each voxel given the previously encoded ones. We important to convey the 3D structure of the PC, but is also then employ the estimated probabilities to code efficiently a block challenging, as the non-regular sampling of point clouds using a context-based arithmetic coder. Our context has variable size and can expand beyond the current block to learn more makes it difficult to use conventional signal processing and accurate probabilities. We also consider using data augmentation compression tools. In this paper, we focus on lossless coding techniques to increase the generalization capability of the learned of point cloud geometry. probability models, in particular in the presence of noise and In particular, we consider the case of voxelized point clouds.
    [Show full text]
  • Integrated Lossy, Near-Lossless, and Lossless Compression of Medical Volumetric Data
    Integrated Lossy, Near-lossless, and Lossless Compression of Medical Volumetric Data Sehoon Yeaa, Sungdae Chob and William A.Pearlmana aCenter for Image Processing Research Electrical, Computer & Systems Engineering Dept. Rensselaer Polytechnic Institute 110 8th St. Troy, NY 12180-3590 bSamsung Electronics, Suwon, South Korea ABSTRACT We propose an integrated, wavelet based, two-stage coding scheme for lossy, near-lossless and lossless compression of medical volumetric data. The method presented determines the bit-rate while encoding for the lossy layer and without any iteration. It is in the spirit of \lossy plus residual coding" and consists of a wavelet-based lossy layer followed by an arithmetic coding of the quantized residual to guarantee a given pixel-wise maximum error bound. We focus on the selection of the optimum bit rate for the lossy coder to achieve the minimum total (lossy plus residual) bit rate in the near-lossless and the lossless cases. We propose a simple and practical method to estimate online the optimal bit rate and provide a theoretical justification for it. Experimental results show that the proposed scheme provides improved, embedded lossy, and lossless performance competitive with the best results published so far in the literature, with an added feature of near-lossless coding. Keywords: near-lossless coding, SPIHT, lossy-to-lossless, medical image 1. INTRODUCTION Most of today's medical imaging techniques produce a huge amount of 3D data. Examples include Magnetic Res- onance(MR),Computerized Tomography(CT), Position Emission Tomography(PET) and 3D ultrasound. Con- siderable research efforts have been made for their efficient compression to facilitate storage and transmission.
    [Show full text]