A Perceptually Adaptive JPEG Coder

Total Page:16

File Type:pdf, Size:1020Kb

A Perceptually Adaptive JPEG Coder A perceptually adaptive JPEG coder Henry Hoi-Yu Tong A thesis submitted in conformity with the requirernents for the degree of Master of Applied Science. Graduate Department of Electrical and Computer Engineering, University of Toronto @Copyright by Henry Tong 1997 National Library Bibliothéque nationale 1*1 of Canada du Canada Acquisitions and Acquisitions et Bibliographie Services services bibliographiques 395 Wellington Street 395. rue Wellington OttawaON K1AON4 OttawaON K1AON4 Canada Canada The author has granted a non- L'auteur a accordé une licence non exclusive licence aiiowing the exclusive permettant à la National Library of Canada to Bibliothèque nationale du Canada de reproduce, loan, distribute or seli reproduire, prêter, distribuer ou copies of this thesis in microfom, vendre des copies de cette thèse sous paper or electronic formats. la forme de microfiche/film, de reproduction sur papier ou sur format électronique. The author retains ownership of the L'auteur conserve la propriété du copyright in this thesis. Neither the droit d'auteur qui protège cette thèse. thesis nor substantid extracts fiom it Ni la thèse ni des extraits substantiels may be printed or otherwise de celle-ci ne doivent être imprimés reproduced without the author's ou autrement reproduits sans son permission. autorisation. A perceptually adaptive JPEG coder A thesis for the Dcgree of' 3Iaster of Applied Science, 1997 b'- Henry Hoi-Yu Tong Department of Electrical and Cornputer Engineering University of Toronto Abstract A pcrceptually adaptive -JPEG coder is irnplernented in this research. The major part of the implementation involves the design of a perceptual model that is based on the texture masking and luminance masking properties of the human visual system (HVS). The main objective is to compute a local multiplier map that can be used to scale the quantization matris (QM) so that fewer bits are used to represent the perceptually less important areas of an image. The texture masking model is based on a proposed block classification algorithm to differentiate between the plain. edge. and texture blocks. An adaptive luminance masking scherne is also proposed n-hich adaptively adjusts the luminance masking strategy depending on an image's mean luminance value. Experimental results show that the adaptive coder provides savings in bit-rate over baseline JPEG. with no overall loss in perceptual quality according to a subject test. Acknowledgment s It is my pleasure to acknowledge and thank al1 people who have helped rnc through- out this researcli. 1 would like to thank rny supervisor, Prof. .A. N. Venetsanopoulos for his continucd support and guidance. Special thanks are due to Kostas Plataniotis for providing invaluable suggestions and ideas that have helpctl so much. I am also grateful to Huawu Lino Ido Rabinovitch, Lowe11 Winger, and Xicos Herodotou for the estremelÿ useful advicc, suggestions, and the friendly conversations we had during the long hours in the DSP lab. I would like to thank Prof. Iiatzela, Prof. Lee, and Prof. Sousa for scrving in the oral examination conimittee. 1 would also likc to thank Sicos and Lyriral Peart for helping me prepare for the examinat,ion. and Dimitrios Androiitsos and Kostas for proofreading the t hesis. Special thanks are also owed to al1 my friends in the communications group and the university for offering encouragement, helpful suggestions. and friendly chats. I would also like to acknowledge the Independent JPEG Group for providing the .JPEG software that forms the bais source code for the research. 1 wish to thank the following individuals for participating in the subjective testing and providing me with feedback and useful comments that were invaluable in com- pleting the perceptual aspect of the thesis: Changshan Xiao. Sheldon Tarn, Wing Tarn, Lowell, Michael Yu , Kostas, Anestis Karasaridis, Peter Androutsos. Evan' Wai Ming Louie. Liiian Herodotou, Stanley Ma, Kevin Gior, Halim Yanikomeroglu. Chao Zeng, Yimen, Dimitrios, Karen Cheung, and Ido. Last but certainly not least, my special thanks to my family for their encourage- ment, patience and understanding, despite the many long distance phone calls that 1 missed during the course of this research. Contents 1 Introduction 1 1.1 Rcsearch objectives and thesis oiitline .................. 3. 2 Review of image compression and some human visual system prop- erties 4 2.1 Classical information t heory ....................... 1 2.2 Lossless and lossy image compression .................. 8 2.3 Relevant human viaual system properties ................ 12 2.3.1 Colour space and irnage coding ................. 12 23.2 The filtering and masking properties .............. 14 3 The JPEG image compression standard 19 3.1 The transform coding mode1 ....................... 21 3.1.1 Tracsformation .......................... 21 3.1.2 Quantization ........................... 37-- 3.1.3 Entropy coding .......................... 26 4 Design of the adaptive JPEG coder 28 4.1 Perceptually adaptive JPEG coding ................... 28 4.1. The two adaptive coding modes ................. 30 4.1.1.1 The JPEG compatibility mode ............ 30 4.1.1.2 The overhead mode .................. 31 4.1.1.3 The difference between the two adaptive modes ... 33 4.2 Design of the perceptrial mode1 ..................... 3.5 4.2.1 Linear threshold elevation modeling ............... 36 4.2.2 The testure rnasking mode1 ................ .. 36 4.3.2.1 Block classification ................... 37 4.2.2.2 Calculation of the texture masking factor ...... 45 -4.2.3 The luminance niasking mode1 .................. 17 423.1 The luminance sensitivity subjective esperiment ... 47 1.2.3.2 The influence of the viewing conditions ........ 19 4.2.3.3 Calculation of the luminance masking factor ..... 50 4.2.4 Processing for the subsampled chrominance channels ..... 34 - - 4.3 Coding of the overhead information ................... 35 5 Experimentd results 58 5.1 Introduction and set-up of the coder parameters ............ 58 5.2 -4 coding esample with the image lenna ................ 62 5.3 Compression results for a fised quality factor .............. 65 5.3.1 Objective results ......................... 65 3.3.2 Subjective results ......................... 70 5.3.2.1 Description of the subjective test ........... 70 5.3.2.2 Discussions ....................... 73 5.4 Compression results for varying quality factors ............. 53 5.4.1 Objective results ......................... 83 5-42 Subjective results ......................... 85 5.5 Computational complexity of the coder ................. 88 6 Conclusions 92 6 Future work ................................ 97 A Possible applications for adapt ive JPEG coding 101 Bibliography List of Figures 2.1 The rate-distortion function ....................... 6 2.2 Classification of image compression methods .............. 8 2.3 Loszless and lossy compression general frarnework ........... 3 '2.1 The modulation transfer function .................... 1.5 2.5 Background luminance and the visibilit- t hreshold ........... 16 2.6 Distortion risibilit~. YS background brïghtness ............. 17 7.7 Texture (spatial j masking effects .................... 17 3.1 .i baseline .lPEG compression sysrem ......-.--....-... 20 3-2 The .JPEG luminance quantization marris ............... 2.5 3.3 The .l PEG chrominance quantization marris .............. 2.5 3.4 The QII multipiying factor as a function of the .JPEC. qualit'- factor . 26 3.5 The zig-zag ordering of AC coefficients ................. 27 1.1 The baseline JPEG-compatible encoder ................. 31 1.1 The overhead mode encoder ....................... 3'2 1.3 JPEG compatibility mode rs overhead mode .............. 34 1.1 Linear threshold elevation modeling ................... 36 4.5 The block classification indicative areas ................. 38 4.6 8 x 8 blocks from the image sail .................... 39 4.7 The block classification algorithm .................... 40 4.8 Locally adaptive cclassification correction ................ 12 4.9 la) barbara . (b) block classification ................... 13 1.10 (a) peppers . (b) block classification ................... 13 4.1 1 (a) .si~sie. (11) hlock classification .................... 44 1.12 (a) horises . (b) block classificar ion ................... 14 4.13 The luminance masking sihjectiw eelprrirncnt ............. 47 4.11 The noise visibility thressholds vs . background luminance ....... 48 4.15 Luminance masking linear modeling ................... 52 4.16 -4 16 x 16 area: four Y biocks . one Cr block . one Ch block ...... 54 - - 4-17 Generation of the chrominancr multiplier ................ JJ 5 (a) lenna . (b) harham. (c) bouts. Id) gofdhill. (e) harbor ........ 61 5 .2 (a) bridge . (b) peppers . (cl mandrill. (ci) fIowes. (e) lad!/ ....... 61 5.3 (a) sail . ibj hats. (cj sirrdow.~. (ri! houses. le) pl........... 61 .J .4 la)nl;er.~h)motorhike.(c)lock.................... 61 - - 3.3 The original lenna image ......................... 63 5.6 Block classification of the lenna image ................. 63 - - 3.r The texture masking multiplier map for lenna ............. 61 5.8 The luminance masking multiplier map for lenna ........... 61 3.9 The combined multiplier rnap for lenna ................. 66 5.10 The reconsrructed lennn image . JPEG cornpatibility mode ...... 66 5-11 Bit-rate savings vs . badine JPEG bit-rate (colour images) ...... 68 -- 5-12 The baseline JPEG coded ienna image ................. 1, 5-13 The A-JPEG coded lenna image - JPEG comparibility mode .... 78 5-14
Recommended publications
  • Melinda's Marks Merit Main Mantle SYDNEY STRIDERS
    SYDNEY STRIDERS ROAD RUNNERS’ CLUB AUSTRALIA EDITION No 108 MAY - AUGUST 2009 Melinda’s marks merit main mantle This is proving a “best-so- she attained through far” year for Melinda. To swimming conflicted with date she has the fastest her transition to running. time in Australia over 3000m. With a smart 2nd Like all top runners she at the State Open 5000m does well over 100k a champs, followed by a week in training, win at the State Open 10k consisting of a variety of Road Champs, another sessions: steady pace, win at the Herald Half medium pace, long slow which doubles as the runs, track work, fartlek, State Half Champs and a hills, gym work and win at the State Cross swimming! country Champs, our Melinda is looking like Springs under her shoes give hot property. Melinda extra lift Melinda began her sports Continued Page 3 career as a swimmer. By 9 years of age she was representing her club at State level. She held numerous records for INSIDE BLISTER 108 Breaststroke and Lisa facing racing pacing Butterfly. Her switch to running came after the McKinney makes most of death of her favourite marvellous mud moment Coach and because she Weather woe means Mo wasn’t growing as big as can’t crow though not slow! her fellow competitors. She managed some pretty fast times at inter-schools Brent takes tumble at Trevi champs and Cross Country before making an impression in the Open category where she has Champion Charles cheered steadily improved. by chance & chase challenge N’Lotsa Uthastuff Melinda credits her swimming background for endurance
    [Show full text]
  • The Basic Principles of Data Compression
    The Basic Principles of Data Compression Author: Conrad Chung, 2BrightSparks Introduction Internet users who download or upload files from/to the web, or use email to send or receive attachments will most likely have encountered files in compressed format. In this topic we will cover how compression works, the advantages and disadvantages of compression, as well as types of compression. What is Compression? Compression is the process of encoding data more efficiently to achieve a reduction in file size. One type of compression available is referred to as lossless compression. This means the compressed file will be restored exactly to its original state with no loss of data during the decompression process. This is essential to data compression as the file would be corrupted and unusable should data be lost. Another compression category which will not be covered in this article is “lossy” compression often used in multimedia files for music and images and where data is discarded. Lossless compression algorithms use statistic modeling techniques to reduce repetitive information in a file. Some of the methods may include removal of spacing characters, representing a string of repeated characters with a single character or replacing recurring characters with smaller bit sequences. Advantages/Disadvantages of Compression Compression of files offer many advantages. When compressed, the quantity of bits used to store the information is reduced. Files that are smaller in size will result in shorter transmission times when they are transferred on the Internet. Compressed files also take up less storage space. File compression can zip up several small files into a single file for more convenient email transmission.
    [Show full text]
  • Source Coding: Part I of Fundamentals of Source and Video Coding
    Foundations and Trends R in sample Vol. 1, No 1 (2011) 1–217 c 2011 Thomas Wiegand and Heiko Schwarz DOI: xxxxxx Source Coding: Part I of Fundamentals of Source and Video Coding Thomas Wiegand1 and Heiko Schwarz2 1 Berlin Institute of Technology and Fraunhofer Institute for Telecommunica- tions — Heinrich Hertz Institute, Germany, [email protected] 2 Fraunhofer Institute for Telecommunications — Heinrich Hertz Institute, Germany, [email protected] Abstract Digital media technologies have become an integral part of the way we create, communicate, and consume information. At the core of these technologies are source coding methods that are described in this text. Based on the fundamentals of information and rate distortion theory, the most relevant techniques used in source coding algorithms are de- scribed: entropy coding, quantization as well as predictive and trans- form coding. The emphasis is put onto algorithms that are also used in video coding, which will be described in the other text of this two-part monograph. To our families Contents 1 Introduction 1 1.1 The Communication Problem 3 1.2 Scope and Overview of the Text 4 1.3 The Source Coding Principle 5 2 Random Processes 7 2.1 Probability 8 2.2 Random Variables 9 2.2.1 Continuous Random Variables 10 2.2.2 Discrete Random Variables 11 2.2.3 Expectation 13 2.3 Random Processes 14 2.3.1 Markov Processes 16 2.3.2 Gaussian Processes 18 2.3.3 Gauss-Markov Processes 18 2.4 Summary of Random Processes 19 i ii Contents 3 Lossless Source Coding 20 3.1 Classification
    [Show full text]
  • Optimal Parsing for Dictionary Text Compression Alessio Langiu
    Optimal Parsing for dictionary text compression Alessio Langiu To cite this version: Alessio Langiu. Optimal Parsing for dictionary text compression. Other [cs.OH]. Université Paris-Est, 2012. English. NNT : 2012PEST1091. tel-00804215 HAL Id: tel-00804215 https://tel.archives-ouvertes.fr/tel-00804215 Submitted on 25 Mar 2013 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. UniversitadegliStudidiPalermo` Dipartimento di Matematica e Applicazioni Dottorato di Ricerca in Matematica e Informatica XXIII◦ Ciclo - S.S.D. Inf/01 UniversiteParis-Est´ Ecole´ doctorale MSTIC Th´ese de doctorat en Informatique Optimal Parsing for Dictionary Text Compression Author: LANGIU Alessio Ph.D. Thesis in Computer Science April 3, 2012 PhD Commission: Thesis Directors: Prof. CROCHEMORE Maxime University of Paris-Est Prof. RESTIVO Antonio University of Palermo Examiners: Prof. ILIOPOULOS Costas King’s College London Prof. LECROQ Thierry University of Rouen Prof. MIGNOSI Filippo University of L’Aquila Referees: Prof. GROSSI Roberto University of Pisa Prof. ILIOPOULOS Costas King’s College London Prof. LECROQ Thierry University of Rouen ii Summary Dictionary-based compression algorithms include a parsing strategy to transform the input text into a sequence of dictionary phrases. Given a text, such process usually is not unique and, for compression purposes, it makes sense to find one of the possible parsing that minimize the final compres- sion ratio.
    [Show full text]
  • Acronis Revive 2019
    Acronis Revive 2019 Table of contents 1 Introduction to Acronis Revive 2019 .................................................................................3 1.1 Acronis Revive 2019 Features .................................................................................................... 3 1.2 System Requirements and Installation Notes ........................................................................... 4 1.3 Technical Support ...................................................................................................................... 5 2 Data Recovery Using Acronis Revive 2019 .........................................................................6 2.1 Recover Lost Files from Existing Logical Disks ........................................................................... 7 2.1.1 Searching for a File ........................................................................................................................................ 16 2.1.2 Finding Previous File Versions ...................................................................................................................... 18 2.1.3 File Masks....................................................................................................................................................... 19 2.1.4 Regular Expressions ...................................................................................................................................... 20 2.1.5 Previewing Files ............................................................................................................................................
    [Show full text]
  • Archive and Compressed [Edit]
    Archive and compressed [edit] Main article: List of archive formats • .?Q? – files compressed by the SQ program • 7z – 7-Zip compressed file • AAC – Advanced Audio Coding • ace – ACE compressed file • ALZ – ALZip compressed file • APK – Applications installable on Android • AT3 – Sony's UMD Data compression • .bke – BackupEarth.com Data compression • ARC • ARJ – ARJ compressed file • BA – Scifer Archive (.ba), Scifer External Archive Type • big – Special file compression format used by Electronic Arts for compressing the data for many of EA's games • BIK (.bik) – Bink Video file. A video compression system developed by RAD Game Tools • BKF (.bkf) – Microsoft backup created by NTBACKUP.EXE • bzip2 – (.bz2) • bld - Skyscraper Simulator Building • c4 – JEDMICS image files, a DOD system • cab – Microsoft Cabinet • cals – JEDMICS image files, a DOD system • cpt/sea – Compact Pro (Macintosh) • DAA – Closed-format, Windows-only compressed disk image • deb – Debian Linux install package • DMG – an Apple compressed/encrypted format • DDZ – a file which can only be used by the "daydreamer engine" created by "fever-dreamer", a program similar to RAGS, it's mainly used to make somewhat short games. • DPE – Package of AVE documents made with Aquafadas digital publishing tools. • EEA – An encrypted CAB, ostensibly for protecting email attachments • .egg – Alzip Egg Edition compressed file • EGT (.egt) – EGT Universal Document also used to create compressed cabinet files replaces .ecab • ECAB (.ECAB, .ezip) – EGT Compressed Folder used in advanced systems to compress entire system folders, replaced by EGT Universal Document • ESS (.ess) – EGT SmartSense File, detects files compressed using the EGT compression system. • GHO (.gho, .ghs) – Norton Ghost • gzip (.gz) – Compressed file • IPG (.ipg) – Format in which Apple Inc.
    [Show full text]
  • The Pillars of Lossless Compression Algorithms a Road Map and Genealogy Tree
    International Journal of Applied Engineering Research ISSN 0973-4562 Volume 13, Number 6 (2018) pp. 3296-3414 © Research India Publications. http://www.ripublication.com The Pillars of Lossless Compression Algorithms a Road Map and Genealogy Tree Evon Abu-Taieh, PhD Information System Technology Faculty, The University of Jordan, Aqaba, Jordan. Abstract tree is presented in the last section of the paper after presenting the 12 main compression algorithms each with a practical This paper presents the pillars of lossless compression example. algorithms, methods and techniques. The paper counted more than 40 compression algorithms. Although each algorithm is The paper first introduces Shannon–Fano code showing its an independent in its own right, still; these algorithms relation to Shannon (1948), Huffman coding (1952), FANO interrelate genealogically and chronologically. The paper then (1949), Run Length Encoding (1967), Peter's Version (1963), presents the genealogy tree suggested by researcher. The tree Enumerative Coding (1973), LIFO (1976), FiFO Pasco (1976), shows the interrelationships between the 40 algorithms. Also, Stream (1979), P-Based FIFO (1981). Two examples are to be the tree showed the chronological order the algorithms came to presented one for Shannon-Fano Code and the other is for life. The time relation shows the cooperation among the Arithmetic Coding. Next, Huffman code is to be presented scientific society and how the amended each other's work. The with simulation example and algorithm. The third is Lempel- paper presents the 12 pillars researched in this paper, and a Ziv-Welch (LZW) Algorithm which hatched more than 24 comparison table is to be developed.
    [Show full text]
  • Vital and Health Statistics; Series 10, No. 40
    NATIONAL CENTER For HEALTH STATISTICS VITAL and HEALTH STATISTICS DATA FROM THE NATIONAL HEALTH SURVEY PROPEfW OFTHE PUBLICATIONSBRANCH EOIT(IRIAL LIBRARY characteristics of Persons With Diabetes United States =July 1964-June 1965 Statistics on the prevalence of diabetes as reported in inter- views, the socioeconomic and health characteristics of dia­ betics, their disabi Iity from diobetes and from all chronic con- ditions, and type of medical and hame care. Based on data collected in household interviews during the period July 1964- June 1965. Washington, D.C. Ocfober 1967 U.S. DEPARTMENT OF HEALTH, EDUCATION, AND WELFARE Public Health Service John W. Gardner William H. Stewart Secretary Surgeon General ~ublic Health Service Publication No. 10@O-Series IO-NO. 40 For sale by the Superintendent of Documents, U.S. flovernment F’rinting Office. \\’ashington, D.C. , !20402 - Price 35 cents NATIONAL CENTER FOR HEALTH STATISTICS FORREST E. LIND13R, PH. D., DirectoT THEODORE D. WOOLSEY, Deputy Dzrector OSWALD K. SAGEN, PH. D., Assistant Director /or State Relatzons WALT R. SIMMONS,M.A., Statzstt cal A dvz sor PHILIP S. LAWRENCE, .SC. D. , Pbzning o//zcer ALICE M. WATERHOUSE, M.D., MedrcaI Consultant JAMES E. KELLY, D.D.S., Dental Advzsor LOUIS R. STOLCIS, M.A., Executzve O//7cer DONALD GREEN, in/ormation O//icer DIVISION OF HEALTH INTERVIEW STATISTICS ELIJAH L. WHITE, AM., Dz~ecfo7 ROBERT R. FUCHSBERG, Chief Survey Methods Branch GERALDINE A. GLEESON, Chic/, Analysis and Reports Branch COOPERATION OF THE BUREAU OF THE CENSUS Under the legislation establishing the National Health Survey, the Public Health Service is authorized to use, insofar as possible, the services or facilities of other Federal, State, or private agencies.
    [Show full text]
  • Download Special Issue
    Wireless Communications and Mobile Computing Error Control Codes for Next-Generation Communication Systems: Opportunities and Challenges Lead Guest Editor: Zesong Fei Guest Editors: Jinhong Yuan and Qin Huang Error Control Codes for Next-Generation Communication Systems: Opportunities and Challenges Wireless Communications and Mobile Computing Error Control Codes for Next-Generation Communication Systems: Opportunities and Challenges Lead Guest Editor: Zesong Fei Guest Editors: Jinhong Yuan and Qin Huang Copyright © 2018 Hindawi. All rights reserved. This is a special issue published in “Wireless Communications and Mobile Computing.” All articles are open access articles distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, pro- vided the original work is properly cited. Editorial Board Javier Aguiar, Spain Oscar Esparza, Spain Maode Ma, Singapore Ghufran Ahmed, Pakistan Maria Fazio, Italy Imadeldin Mahgoub, USA Wessam Ajib, Canada Mauro Femminella, Italy Pietro Manzoni, Spain Muhammad Alam, China Manuel Fernandez-Veiga, Spain Álvaro Marco, Spain Eva Antonino-Daviu, Spain Gianluigi Ferrari, Italy Gustavo Marfia, Italy Shlomi Arnon, Israel Ilario Filippini, Italy Francisco J. Martinez, Spain Leyre Azpilicueta, Mexico Jesus Fontecha, Spain Davide Mattera, Italy Paolo Barsocchi, Italy Luca Foschini, Italy Michael McGuire, Canada Alessandro Bazzi, Italy A. G. Fragkiadakis, Greece Nathalie Mitton, France Zdenek Becvar, Czech Republic Sabrina Gaito, Italy Klaus Moessner, UK Francesco Benedetto, Italy Óscar García, Spain Antonella Molinaro, Italy Olivier Berder, France Manuel García Sánchez, Spain Simone Morosi, Italy Ana M. Bernardos, Spain L. J. García Villalba, Spain Kumudu S. Munasinghe, Australia Mauro Biagi, Italy José A. García-Naya, Spain Enrico Natalizio, France Dario Bruneo, Italy Miguel Garcia-Pineda, Spain Keivan Navaie, UK Jun Cai, Canada A.-J.
    [Show full text]
  • The Deep Learning Solutions on Lossless Compression Methods for Alleviating Data Load on Iot Nodes in Smart Cities
    sensors Article The Deep Learning Solutions on Lossless Compression Methods for Alleviating Data Load on IoT Nodes in Smart Cities Ammar Nasif *, Zulaiha Ali Othman and Nor Samsiah Sani Center for Artificial Intelligence Technology (CAIT), Faculty of Information Science & Technology, University Kebangsaan Malaysia, Bangi 43600, Malaysia; [email protected] (Z.A.O.); [email protected] (N.S.S.) * Correspondence: [email protected] Abstract: Networking is crucial for smart city projects nowadays, as it offers an environment where people and things are connected. This paper presents a chronology of factors on the development of smart cities, including IoT technologies as network infrastructure. Increasing IoT nodes leads to increasing data flow, which is a potential source of failure for IoT networks. The biggest challenge of IoT networks is that the IoT may have insufficient memory to handle all transaction data within the IoT network. We aim in this paper to propose a potential compression method for reducing IoT network data traffic. Therefore, we investigate various lossless compression algorithms, such as entropy or dictionary-based algorithms, and general compression methods to determine which algorithm or method adheres to the IoT specifications. Furthermore, this study conducts compression experiments using entropy (Huffman, Adaptive Huffman) and Dictionary (LZ77, LZ78) as well as five different types of datasets of the IoT data traffic. Though the above algorithms can alleviate the IoT data traffic, adaptive Huffman gave the best compression algorithm. Therefore, in this paper, Citation: Nasif, A.; Othman, Z.A.; we aim to propose a conceptual compression method for IoT data traffic by improving an adaptive Sani, N.S.
    [Show full text]
  • (12) United States Patent (10) Patent No.: US 8,013,862 B2 Mathe (45) Date of Patent: Sep
    US0080 13862B2 (12) United States Patent (10) Patent No.: US 8,013,862 B2 Mathe (45) Date of Patent: Sep. 6, 2011 (54) TEXTURE CODEC 2004/O151372 A1* 8, 2004 Reshetov et al. ............. 382,166 2004/0227552 Al 11/2004 Markovic et al. 327/215 (75) Inventor: Zsolt Mathe, Issaquah, WA (US) 2005/0206784 A1* 9, 2005 Li et al. ......... ... 348/441 2006, OO17722 A1 1/2006 Hong et al. ... 345,419 2006/0092.169 A1 5, 2006 Wetzel ....... 345,582 (73) Assignee: Microsoft Corporation, Redmond, WA 2006/0268990 A1* 11, 2006 Lin et al. .... 375,240.24 (US) 2007/0070.083 A1* 3, 2007 Fouladi et al. ................ 345 604 2007, 0140554 A1* 6, 2007 Strom et al. .. ... 382,166 (*) Notice: Subject to any disclaimer, the term of this 2007/0172120 A1, 7/2007 Roimela et al... ... 382,166 patent is extended or adjusted under 35 2007/024208.6 A1* 10/2007 Tsujimoto et al. ............ 345,632 U.S.C. 154(b) by 890 days. OTHER PUBLICATIONS (21) Appl. No.: 11/941,337 DXTex Tool, (C) 2007. http://msdn.microsoft.com/archive, 3 pages, date unknown. (22) Filed: Nov. 16, 2007 van Waveren, J.M.P., “Real-Time DXT Compression'. May 20, 2006, http://cache-www.intel.com, 43 pages. (65) Prior Publication Data van Waveren, J.M.P., “Real-Time Texture Streaming & Decompres sion”, Nov. 11, 2006, http://softwarecommunity.intel.com, 57 pages. US 2009/O128576A1 May 21, 2009 * cited by examiner (51) Int. Cl. G09G 5/00 (2006.01) Primary Examiner — M Good Johnson (52) U.S. Cl.
    [Show full text]
  • Answers to Exercises
    Answers to Exercises A bird does not sing because he has an answer, he sings because he has a song. —Chinese Proverb Intro.1: abstemious, abstentious, adventitious, annelidous, arsenious, arterious, face- tious, sacrilegious. Intro.2: When a software house has a popular product they tend to come up with new versions. A user can update an old version to a new one, and the update usually comes as a compressed file on a floppy disk. Over time the updates get bigger and, at a certain point, an update may not fit on a single floppy. This is why good compression is important in the case of software updates. The time it takes to compress and decompress the update is unimportant since these operations are typically done just once. Recently, software makers have taken to providing updates over the Internet, but even in such cases it is important to have small files because of the download times involved. 1.1: (1) ask a question, (2) absolutely necessary, (3) advance warning, (4) boiling hot, (5) climb up, (6) close scrutiny, (7) exactly the same, (8) free gift, (9) hot water heater, (10) my personal opinion, (11) newborn baby, (12) postponed until later, (13) unexpected surprise, (14) unsolved mysteries. 1.2: A reasonable way to use them is to code the five most-common strings in the text. Because irreversible text compression is a special-purpose method, the user may know what strings are common in any particular text to be compressed. The user may specify five such strings to the encoder, and they should also be written at the start of the output stream, for the decoder’s use.
    [Show full text]