JPEG Image Compression Using the Discrete Cosine Transform: an Overview, Applications, and Hardware Implementation Ahmad Shawahna, Md
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Information and Compression Introduction Summary of Course
Information and compression Introduction ■ We will be looking at what information is Information and Compression ■ Methods of measuring it ■ The idea of compression ■ 2 distinct ways of compressing images Paul Cockshott Faraday partnership Summary of Course Agenda ■ At the end of this lecture you should have ■ Encoding systems an idea what information is, its relation to ■ BCD versus decimal order, and why data compression is possible ■ Information efficiency ■ Shannons Measure ■ Chaitin, Kolmogorov Complexity Theory ■ Relation to Goedels theorem Overview Connections ■ Data coms channels bounded ■ Simple encoding example shows different ■ Images take great deal of information in raw densities possible form ■ Shannons information theory deals with ■ Efficient transmission requires denser transmission encoding ■ Chaitin relates this to Turing machines and ■ This requires that we understand what computability information actually is ■ Goedels theorem limits what we know about compression paul 1 Information and compression Vocabulary What is information ■ information, ■ Information measured in bits ■ entropy, ■ Bit equivalent to binary digit ■ compression, ■ Why use binary system not decimal? ■ redundancy, ■ Decimal would seem more natural ■ lossless encoding, ■ Alternatively why not use e, base of natural ■ lossy encoding logarithms instead of 2 or 10 ? Voltage encoding Noise Immunity Binary voltage ■ ■ ■ 5v Digital information BINARY DECIMAL encoded as ranges of ■ 2 values 0,1 ■ 10 values 0..9 continuous variables 1 ■ 1 isolation band ■ -
An Improved Fast Fractal Image Compression Coding Method
Rev. Téc. Ing. Univ. Zulia. Vol. 39, Nº 9, 241 - 247, 2016 doi:10.21311/001.39.9.32 An Improved Fast Fractal Image Compression Coding Method Haibo Gao, Wenjuan Zeng, Jifeng Chen, Cheng Zhang College of Information Science and Engineering, Hunan International Economics University, Changsha 410205, Hunan, China Abstract The main purpose of image compression is to use as many bits as possible to represent the source image by eliminating the image redundancy with acceptable restoration. Fractal image compression coding has achieved good results in the field of image compression due to its high compression ratio, fast decoding speed as well as independency between decoding image and resolution. However, there is a big problem in this method, i.e. long coding time because there are considerable searches of fractal blocks in the coding. Based on the theoretical foundation of fractal coding and fractal compression algorithm, this paper researches every step of this compression method, including block partition, block search and the representation of affine transformation coefficient, in order to come up with optimization and improvement strategies. The simulation result shows that the algorithm of this paper can achieve excellent effects in coding time and compression ratio while ensuring better image quality. The work on fractal coding in this paper has made certain contributions to the research of fractal coding. Key words: Image Compression, Fractal Theory, Coding and Decoding. 1. INTRODUCTION Image compression is aimed to use as many bytes as possible to represent and transmit the original big image and to have a better quality in the restored image. -
Discrete Cosine Transform Based Image Fusion Techniques VPS Naidu MSDF Lab, FMCD, National Aerospace Laboratories, Bangalore, INDIA E.Mail: [email protected]
View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by NAL-IR Journal of Communication, Navigation and Signal Processing (January 2012) Vol. 1, No. 1, pp. 35-45 Discrete Cosine Transform based Image Fusion Techniques VPS Naidu MSDF Lab, FMCD, National Aerospace Laboratories, Bangalore, INDIA E.mail: [email protected] Abstract: Six different types of image fusion algorithms based on 1 discrete cosine transform (DCT) were developed and their , k 1 0 performance was evaluated. Fusion performance is not good while N Where (k ) 1 and using the algorithms with block size less than 8x8 and also the block 1 2 size equivalent to the image size itself. DCTe and DCTmx based , 1 k 1 N 1 1 image fusion algorithms performed well. These algorithms are very N 1 simple and might be suitable for real time applications. 1 , k 0 Keywords: DCT, Contrast measure, Image fusion 2 N 2 (k 1 ) I. INTRODUCTION 2 , 1 k 2 N 2 1 Off late, different image fusion algorithms have been developed N 2 to merge the multiple images into a single image that contain all useful information. Pixel averaging of the source images k 1 & k 2 discrete frequency variables (n1 , n 2 ) pixel index (the images to be fused) is the simplest image fusion technique and it often produces undesirable side effects in the fused image Similarly, the 2D inverse discrete cosine transform is defined including reduced contrast. To overcome this side effects many as: researchers have developed multi resolution [1-3], multi scale [4,5] and statistical signal processing [6,7] based image fusion x(n1 , n 2 ) (k 1 ) (k 2 ) N 1 N 1 techniques. -
Digital Signals
Technical Information Digital Signals 1 1 bit t Part 1 Fundamentals Technical Information Part 1: Fundamentals Part 2: Self-operated Regulators Part 3: Control Valves Part 4: Communication Part 5: Building Automation Part 6: Process Automation Should you have any further questions or suggestions, please do not hesitate to contact us: SAMSON AG Phone (+49 69) 4 00 94 67 V74 / Schulung Telefax (+49 69) 4 00 97 16 Weismüllerstraße 3 E-Mail: [email protected] D-60314 Frankfurt Internet: http://www.samson.de Part 1 ⋅ L150EN Digital Signals Range of values and discretization . 5 Bits and bytes in hexadecimal notation. 7 Digital encoding of information. 8 Advantages of digital signal processing . 10 High interference immunity. 10 Short-time and permanent storage . 11 Flexible processing . 11 Various transmission options . 11 Transmission of digital signals . 12 Bit-parallel transmission. 12 Bit-serial transmission . 12 Appendix A1: Additional Literature. 14 99/12 ⋅ SAMSON AG CONTENTS 3 Fundamentals ⋅ Digital Signals V74/ DKE ⋅ SAMSON AG 4 Part 1 ⋅ L150EN Digital Signals In electronic signal and information processing and transmission, digital technology is increasingly being used because, in various applications, digi- tal signal transmission has many advantages over analog signal transmis- sion. Numerous and very successful applications of digital technology include the continuously growing number of PCs, the communication net- work ISDN as well as the increasing use of digital control stations (Direct Di- gital Control: DDC). Unlike analog technology which uses continuous signals, digital technology continuous or encodes the information into discrete signal states (Fig. 1). When only two discrete signals states are assigned per digital signal, these signals are termed binary si- gnals. -
How I Came up with the Discrete Cosine Transform Nasir Ahmed Electrical and Computer Engineering Department, University of New Mexico, Albuquerque, New Mexico 87131
mxT*L. BImL4L. PRocEsSlNG 1,4-5 (1991) How I Came Up with the Discrete Cosine Transform Nasir Ahmed Electrical and Computer Engineering Department, University of New Mexico, Albuquerque, New Mexico 87131 During the late sixties and early seventies, there to study a “cosine transform” using Chebyshev poly- was a great deal of research activity related to digital nomials of the form orthogonal transforms and their use for image data compression. As such, there were a large number of T,(m) = (l/N)‘/“, m = 1, 2, . , N transforms being introduced with claims of better per- formance relative to others transforms. Such compari- em- lh) h = 1 2 N T,(m) = (2/N)‘%os 2N t ,...) . sons were typically made on a qualitative basis, by viewing a set of “standard” images that had been sub- jected to data compression using transform coding The motivation for looking into such “cosine func- techniques. At the same time, a number of researchers tions” was that they closely resembled KLT basis were doing some excellent work on making compari- functions for a range of values of the correlation coef- sons on a quantitative basis. In particular, researchers ficient p (in the covariance matrix). Further, this at the University of Southern California’s Image Pro- range of values for p was relevant to image data per- cessing Institute (Bill Pratt, Harry Andrews, Ali Ha- taining to a variety of applications. bibi, and others) and the University of California at Much to my disappointment, NSF did not fund the Los Angeles (Judea Pearl) played a key role. -
Hardware Architecture for Elliptic Curve Cryptography and Lossless Data Compression
Hardware architecture for elliptic curve cryptography and lossless data compression by Miguel Morales-Sandoval A thesis presented to the National Institute for Astrophysics, Optics and Electronics in partial ful¯lment of the requirement for the degree of Master of Computer Science Thesis Advisor: PhD. Claudia Feregrino-Uribe Computer Science Department National Institute for Astrophysics, Optics and Electronics Tonantzintla, Puebla M¶exico December 2004 Abstract Data compression and cryptography play an important role when transmitting data across a public computer network. While compression reduces the amount of data to be transferred or stored, cryptography ensures that data is transmitted with reliability and integrity. Compression and encryption have to be applied in the correct way: data are compressed before they are encrypted. If it were the opposite case the result of the cryptographic operation would be illegible data and no patterns or redundancy would be present, leading to very poor or no compression. In this research work, a hardware architecture that joins lossless compression and public-key cryptography for secure data transmission applications is discussed. The architecture consists of a dictionary-based lossless data compressor to compress the incoming data, and an elliptic curve crypto- graphic module that performs two EC (elliptic curve) cryptographic schemes: encryp- tion and digital signature. For lossless data compression, the dictionary-based LZ77 algorithm is implemented using a systolic array aproach. The elliptic curve cryptosys- tem is de¯ned over the binary ¯eld F2m , using polynomial basis, a±ne coordinates and the binary method to compute an scalar multiplication. While the model of the complete system was implemented in software, the hardware architecture was described in the Very High Speed Integrated Circuit Hardware Description Language (VHDL). -
A Review of Data Compression Technique
International Journal of Computer Science Trends and Technology (IJCST) – Volume 2 Issue 4, Jul-Aug 2014 RESEARCH ARTICLE OPEN ACCESS A Review of Data Compression Technique Ritu1, Puneet Sharma2 Department of Computer Science and Engineering, Hindu College of Engineering, Sonipat Deenbandhu Chhotu Ram University, Murthal Haryana-India ABSTRACT With the growth of technology and the entrance into the Digital Age, the world has found itself amid a vast amount of information. Dealing with such enormous amount of information can often present difficulties. Digital information must be stored and retrieved in an efficient manner, in order for it to be put to practical use. Compression is one way to deal with this problem. Images require substantial storage and transmission resources, thus image compression is advantageous to reduce these requirements. Image compression is a key technology in transmission and storage of digital images because of vast data associated with them. This paper addresses about various image compression techniques. There are two types of image compression: lossless and lossy. Keywords:- Image Compression, JPEG, Discrete wavelet Transform. I. INTRODUCTION redundancies. An inverse process called decompression Image compression is important for many applications (decoding) is applied to the compressed data to get the that involve huge data storage, transmission and retrieval reconstructed image. The objective of compression is to such as for multimedia, documents, videoconferencing, and reduce the number of bits as much as possible, while medical imaging. Uncompressed images require keeping the resolution and the visual quality of the considerable storage capacity and transmission bandwidth. reconstructed image as close to the original image as The objective of image compression technique is to reduce possible. -
Reception Performance Improvement of AM/FM Tuner by Digital Signal Processing Technology
Reception performance improvement of AM/FM tuner by digital signal processing technology Akira Hatakeyama Osamu Keishima Kiyotaka Nakagawa Yoshiaki Inoue Takehiro Sakai Hirokazu Matsunaga Abstract With developments in digital technology, CDs, MDs, DVDs, HDDs and digital media have become the mainstream of car AV products. In terms of broadcasting media, various types of digital broadcasting have begun in countries all over the world. Thus, there is a demand for smaller and thinner products, in order to enhance radio performance and to achieve consolidation with the above-mentioned digital media in limited space. Due to these circumstances, we are attaining such performance enhancement through digital signal processing for AM/FM IF and beyond, and both tuner miniaturization and lighter products have been realized. The digital signal processing tuner which we will introduce was developed with Freescale Semiconductor, Inc. for the 2005 line model. In this paper, we explain regarding the function outline, characteristics, and main tech- nology involved. 22 Reception performance improvement of AM/FM tuner by digital signal processing technology Introduction1. Introduction from IF signals, interference and noise prevention perfor- 1 mance have surpassed those of analog systems. In recent years, CDs, MDs, DVDs, and digital media have become the mainstream in the car AV market. 2.2 Goals of digitalization In terms of broadcast media, with terrestrial digital The following items were the goals in the develop- TV and audio broadcasting, and satellite broadcasting ment of this digital processing platform for radio: having begun in Japan, while overseas DAB (digital audio ①Improvements in performance (differentiation with broadcasting) is used mainly in Europe and SDARS (satel- other companies through software algorithms) lite digital audio radio service) and IBOC (in band on ・Reduction in noise (improvements in AM/FM noise channel) are used in the United States, digital broadcast- reduction performance, and FM multi-pass perfor- ing is expected to increase in the future. -
Next-Generation Photonic Transport Network Using Digital Signal Processing
Next-Generation Photonic Transport Network Using Digital Signal Processing Yasuhiko Aoki Hisao Nakashima Shoichiro Oda Paparao Palacharla Coherent optical-fiber transmission technology using digital signal processing is being actively researched and developed for use in transmitting high-speed signals of the 100-Gb/s class over long distances. It is expected that network capacity can be further expanded by operating a flexible photonic network having high spectral efficiency achieved by applying an optimal modulation format and signal processing algorithm depending on the transmission distance and required bit rate between transmit and receive nodes. A system that uses such adaptive modulation technology should be able to assign in real time a transmission path between a transmitter and receiver. Furthermore, in the research and development stage, optical transmission characteristics for each modulation format and signal processing algorithm to be used by the system should be evaluated under emulated quasi-field conditions. We first discuss how the use of transmitters and receivers supporting multiple modulation formats can affect network capacity. We then introduce an evaluation platform consisting of a coherent receiver based on a field programmable gate array (FPGA) and polarization mode dispersion/polarization dependent loss (PMD/PDL) emulators with a recirculating-loop experimental system. Finally, we report the results of using this platform to evaluate the transmission characteristics of 112-Gb/s dual-polarization, quadrature phase shift keying (DP-QPSK) signals by emulating the factors that degrade signal transmission in a real environment. 1. Introduction amplifiers—which is becoming a limiting factor In the field of optical communications in system capacity—can be efficiently used. -
A Review on Different Lossless Image Compression Techniques
Ilam Parithi.T et. al. /International Journal of Modern Sciences and Engineering Technology (IJMSET) ISSN 2349-3755; Available at https://www.ijmset.com Volume 2, Issue 4, 2015, pp.86-94 A Review on Different Lossless Image Compression Techniques 1Ilam Parithi.T * 2Balasubramanian.R Research Scholar/CSE Professor/CSE Manonmaniam Sundaranar Manonmaniam Sundaranar University, University, Tirunelveli, India Tirunelveli, India [email protected] [email protected] Abstract In the Morden world sharing and storing the data in the form of image is a great challenge for the social networks. People are sharing and storing a millions of images for every second. Though the data compression is done to reduce the amount of space required to store a data, there is need for a perfect image compression algorithm which will reduce the size of data for both sharing and storing. Keywords: Huffman Coding, Run Length Encoding (RLE), Arithmetic Coding, Lempel – Ziv-Welch Coding (LZW), Differential Pulse Code Modulation (DPCM). 1. INTRODUCTION: The Image compression is a technique for effectively coding the digital image by minimizing the numbers of bits needed for representing the image. The aim is to reduce the storage space, transmission cost and to maintain a good quality. The compression is also achieved by removing the redundancies like coding redundancy, inter pixel redundancy, and psycho visual redundancy. Generally the compression techniques are divided into two types. i) Lossy compression techniques and ii) Lossless compression techniques. Compression Techniques Lossless Compression Lossy Compression Run Length Coding Huffman Coding Wavelet based Fractal Compression Compression Arithmetic Coding LZW Coding Vector Quantization Block Truncation Coding Fig. -
Improving Mobile IP Handover Latency on End-To-End TCP in UMTS/WCDMA Networks
Improving Mobile IP Handover Latency on End-to-End TCP in UMTS/WCDMA Networks Chee Kong LAU A thesis submitted in fulfilment of the requirements for the degree of Master of Engineering (Research) in Electrical Engineering School of Electrical Engineering and Telecommunications The University of New South Wales March, 2006 CERTIFICATE OF ORIGINALITY I hereby declare that this submission is my own work and to the best of my knowledge it contains no materials previously published or written by another person, nor material which to a substantial extent has been accepted for the award of any other degree or diploma at UNSW or any other educational institution, except where due acknowledgement is made in the thesis. Any contribution made to the research by others, with whom I have worked at UNSW or elsewhere, is explicitly acknowledged in the thesis. I also declare that the intellectual content of this thesis is the product of my own work, except to the extent that assistance from others in the project's design and conception or in style, presentation and linguistic expression is acknowledged. (Signed)_________________________________ i Dedication To: My girl-friend (Jeslyn) for her love and caring, my parents for their endurance, and to my friends and colleagues for their understanding; because while doing this thesis project I did not manage to accompany them. And Lord Buddha for His great wisdom and compassion, and showing me the actual path to liberation. For this, I take refuge in the Buddha, the Dharma, and the Sangha. ii Acknowledgement There are too many people for their dedications and contributions to make this thesis work a reality; without whom some of the contents of this thesis would not have been realised. -
Digital Signal Processing: Theory and Practice, Hardware and Software
AC 2009-959: DIGITAL SIGNAL PROCESSING: THEORY AND PRACTICE, HARDWARE AND SOFTWARE Wei PAN, Idaho State University Wei Pan is Assistant Professor and Director of VLSI Laboratory, Electrical Engineering Department, Idaho State University. She has several years of industrial experience including Siemens (project engineering/management.) Dr. Pan is an active member of ASEE and IEEE and serves on the membership committee of the IEEE Education Society. S. Hossein Mousavinezhad, Idaho State University S. Hossein Mousavinezhad is Professor and Chair, Electrical Engineering Department, Idaho State University. Dr. Mousavinezhad is active in ASEE and IEEE and is an ABET program evaluator. Hossein is the founding general chair of the IEEE International Conferences on Electro Information Technology. Kenyon Hart, Idaho State University Kenyon Hart is Specialist Engineer and Associate Lecturer, Electrical Engineering Department, Idaho State University, Pocatello, Idaho. Page 14.491.1 Page © American Society for Engineering Education, 2009 Digital Signal Processing, Theory/Practice, HW/SW Abstract Digital Signal Processing (DSP) is a course offered by many Electrical and Computer Engineering (ECE) programs. In our school we offer a senior-level, first-year graduate course with both lecture and laboratory sections. Our experience has shown that some students consider the subject matter to be too theoretical, relying heavily on mathematical concepts and abstraction. There are several visible applications of DSP including: cellular communication systems, digital image processing and biomedical signal processing. Authors have incorporated many examples utilizing software packages including MATLAB/MATHCAD in the course and also used classroom demonstrations to help students visualize some difficult (but important) concepts such as digital filters and their design, various signal transformations, convolution, difference equations modeling, signals/systems classifications and power spectral estimation as well as optimal filters.