Analysis of Video Compression Using DCT

Total Page:16

File Type:pdf, Size:1020Kb

Analysis of Video Compression Using DCT Imperial Journal of Interdisciplinary Research (IJIR) Vol-3, Issue-3, 2017 ISSN: 2454-1362, http://www.onlinejournal.in Analysis of Video Compression using DCT Pranavi Patil1, Sanskruti Patil2 & Harshala Shelke3 & Prof. Anand Sankhe4 1,2,3 Bachelor of Engineering in computer, Mumbai University 4Proffessor, Department of Computer Science & Engineering, Mumbai University Maharashtra, India Abstract: Video is most useful media to represent the where in order to get the best possible compression great information. Videos, images are the most efficiency, it considers certain level of distortion. essential approaches to represent data. Now this year, all the communications are done on such ii. Lossless Compression: media. The central problem for the media is its large The lossless compression technique is generally size. Also this large data contains a lot of redundant focused on the decreasing the compressed output information. The huge usage of digital multimedia video' bit rate without any alteration of the frame. leads to inoperable growth of data flow through The decompressed bit-stream is matching to the various mediums. Now a days, to solve the problem original bit-stream. of bandwidth requirement in communication, multimedia data is a big challenge. The main 2. Video Compression concept in the present paper is about the Discrete The main unit behind the video compression Cosine Transform (DCT) algorithm for compressing approach is the video encoder found at the video's size. Keywords: Compression, DCT, PSNR, transmitter side, which encodes the video that is to be MSE, Redundancy. transmitted in form of bits and also the video decoder positioned at the receiver side, which rebuild the Keywords: Compression, DCT, PSNR, MSE, video in its original form based on the bit sequence Redundancy. given at the encoder. Video compression reduces redundancy and irrelevancy. 1. Introduction Video is an effective manner of entertainment Sources of Redundancy: and communication, now a day. But it needs a vast 1) Spatial storage and transmission bandwidth. For example, A 2) Temporal video of 90 minutes, frame rate of 30 frames per second and 750*570 resolution, it will needs 2.78 Irrelevancy: GB storage. The storage and bandwidth need for this Intra coded frame (I) uncompressed video is very high, compressed size of Inter coded frame (P) such video is upto a certain MB's, then requirement Bi-directional frame (B) of storage can be reduced to a large size. Video compression provides the way to compress the video's size. The videos and images contain a big amount of redundant information. The main procedure in the video compression is about rejecting this redundant information, which is unobservable for the human eyes. In the procedure of video compression, algorithm is put onto the input uncompressed video to generate the output compressed video, which can be efficient for transmission or storage. Video compression decreases the file size, so that Fig.1 MPEG Encoding pattern of video frames compressed video needs an realistic amount of download time. Following are the two types of The video combines the large number of compression methods are: frames. Initially, the video is transformed into the sequence of frames which are nothing but images i. Lossy Compression: sequence. The intra coding method used for The investigation of the study in digital video compression of I frame. compression is conquered by lossy compression, Imperial Journal of Interdisciplinary Research (IJIR) Page 721 Imperial Journal of Interdisciplinary Research (IJIR) Vol-3, Issue-3, 2017 ISSN: 2454-1362, http://www.onlinejournal.in Video Compression Techniques: There is several compression techniques developed for compressing the video efficiently. The mainly used techniques are the Discrete Cosine Transform (DCT) compression and Discrete Wavelet Transform (DWT) compression. In video compression, the video is converted into individual frames and then various techniques of compression are applied on that each frame. The compressed frames are decompressed and again a video could be created from those, so that the output of compressed video is obtained. Also the DCT method, initially converts the video into the various number of frames. Each frame is again divided into the small blocks and then the DCT algorithm is applied to each frame. After applying the DCT algorithm, it converts every pixel value into the frequency domain. This conversion process happened in such a way that the low Fig.2 Video compression system using DCT algorithm frequencies are on the top-left and higher frequencies are on the bottom right. Then the quantization is Video compression system using DCT done, because of quantization the DCT coefficients algorithm To compress the video, initially the video become integers as they have been scaled by a needs to be converted into individual frames and then scaling factor. By applying Inverse Discrete Cosine compression techniques are applied on each frame. Transform (IDCT) algorithm, original input frames To compress the picture, DCT applies to each frame can be reformed. and thus, type of the DCT method is an intra-frame In Discrete wavelet transform (DWT), the compression. Once, compression of all the frames compression techniques directly applied on the frame are done, the sequence of compressed frame forms as a whole (i.e., frame need not be divided into the compressed video, whose size is relatively smaller blocks). The main intention of this smaller than original video. compression algorithms is to store up the frame data Basically in the DCT method, the compression as in small space as possible. occurs in a three step process. Firstly, each frame is divided into small blocks and then DCT is applied on 3. Proposed System each frame, after applying DCT on each frame, it In the several communication systems, the converts the entire pixel values into frequency problem with video is its large size only, So that size domain such that the larger frequency pixels reside compression of video is required to save storage into the bottom-right and the smaller frequency space. The compression techniques mainly searches pixels reside into the top-left places in the matrix. As for the redundancies among the several frames and the eye of human is sensible only to lower frequency also the correlation among the several frames to pixels and thus, the higher frequency pixels are obtain compression of high degree. In the proposed rejected. Secondly, the quantization is applied onto method, the up-down sampling based approach is the obtained matrix such that, the coefficients reject used. Here the system is presenting a Discrete Cosine their values after the decimal point. Depending on Transform (DCT) algorithm to perform the video this, the scaling factor is choosed, so that even after compression with scalability factor. the values later than the decimal point are rejected, the value remains almost the similar. Then, finally the coding technique for compression is applied, after that because of applying the Inverse DCT, the original file can be reconstructed. The algorithm used in these technique takes any number of jpeg video frames and the any size video can be compressed. The DCT algorithm is simplest and balanced, so that, it is quite easy to be implemented when compared to the other compression techniques. 4. Error metrics Compressed video output calculated based on PSNR & Compression Ratio. The Mean Square Error Imperial Journal of Interdisciplinary Research (IJIR) Page 722 Imperial Journal of Interdisciplinary Research (IJIR) Vol-3, Issue-3, 2017 ISSN: 2454-1362, http://www.onlinejournal.in (MSE) and the Peak Signal to Noise Ratio (PSNR) are two error metrics available to measure the [4] Chowdhury, M. Mozammel Hoque, and Amina compression quality of the frame image. To compute Khatun. "Image Compression Using Discrete Wavelet the PSNR, first needs to calculate the mean-square Transform." International Journal of Computer Science (2012). error with the help of following equation: MSE [5] Jasmeet kaur1 Ms. Rohini Sharma2, “ A Combined DWT-DCT approach to perform Video compression base Compression ratio is a measure of the reduction of of Frame Redundancy” , September 2012. detailed coefficient of data. [6] M.A. El-dosuky1 and Wesam Ahmed “Jpeg Image Compression Using Discrete Cosine Transform – A Survey” A.M.Raid1, W.M.Khedr2, April 2014. PSNR block help to calculate the peak signal- to-noise ratio, in decibels, between two frames. This [7] Saraswathy, K., D. Vaithiyanathan, and R. ratio is frequently used to measure the quality among Seshasayanan. "A DCT approximation with Low the original that is uncompressed and a compressed Complexity for image compression." Communications and frames. Signal Processing (ICCSP), 2013 International Conference on. IEEE, 2013. PSNR = ) [8] G.Suresh,“A Low Complex Scalable Spatial If the PSNR is higher, then the compression quality Adjacency ACC-DCT Based Video Compression or reconstruction of the frame is also better. Method”, 2010 Second International conference on Computing, Communication and Networking 5. Conclusion Technologiespp. 4244, 2010. In this paper, we have presented a video [9] K.Saraswathy et al. “A DCT Approximation with Low compression algorithm named as Discrete Cosine Complexity For Image Compression”, (2013). Transform (DCT). The input uncompressed video is transformed into frames and the size of the frames is [10] Gupta, Maneesha, and Amit Kumar Garg. "Analysis converted as per the requirement, then the DCT
Recommended publications
  • MPEG Video in Software: Representation, Transmission, and Playback
    High Speed Networking and Multimedia Computing, IS&T/SPIE Symp. on Elec. Imaging Sci. & Tech., San Jose, CA, February 1994. MPEG Video in Software: Representation, Transmission, and Playback Lawrence A. Rowe, Ketan D. Patel, Brian C Smith, and Kim Liu Computer Science Division - EECS University of California Berkeley, CA 94720 ([email protected]) Abstract A software decoder for MPEG-1 video was integrated into a continuous media playback system that supports synchronized playing of audio and video data stored on a file server. The MPEG-1 video playback system supports forward and backward play at variable speeds and random positioning. Sending and receiving side heuristics are described that adapt to frame drops due to network load and the available decoding capacity of the client workstation. A series of experiments show that the playback system adds a small overhead to the stand alone software decoder and that playback is smooth when all frames or very few frames can be decoded. Between these extremes, the system behaves reasonably but can still be improved. 1.0 Introduction As processor speed increases, real-time software decoding of compressed video is possible. We developed a portable software MPEG-1 video decoder that can play small-sized videos (e.g., 160 x 120) in real-time and medium-sized videos within a factor of two of real-time on current workstations [1]. We also developed a system to deliver and play synchronized continuous media streams (e.g., audio, video, images, animation, etc.) on a network [2].Initially, this system supported 8kHz 8-bit audio and hardware-assisted motion JPEG compressed video streams.
    [Show full text]
  • Avid Supported Video File Formats
    Avid Supported Video File Formats 04.07.2021 Page 1 Avid Supported Video File Formats 4/7/2021 Table of Contents Common Industry Formats ............................................................................................................................................................................................................................................................................................................................................................................................... 4 Application & Device-Generated Formats .................................................................................................................................................................................................................................................................................................................................................................. 8 Stereoscopic 3D Video Formats ...................................................................................................................................................................................................................................................................................................................................................................................... 11 Quick Lookup of Common File Formats ARRI..............................................................................................................................................................................................................................................................................................................................................................4
    [Show full text]
  • Video Codec Requirements and Evaluation Methodology
    Video Codec Requirements 47pt 30pt and Evaluation Methodology Color::white : LT Medium Font to be used by customers and : Arial www.huawei.com draft-filippov-netvc-requirements-01 Alexey Filippov, Huawei Technologies 35pt Contents Font to be used by customers and partners : • An overview of applications • Requirements 18pt • Evaluation methodology Font to be used by customers • Conclusions and partners : Slide 2 Page 2 35pt Applications Font to be used by customers and partners : • Internet Protocol Television (IPTV) • Video conferencing 18pt • Video sharing Font to be used by customers • Screencasting and partners : • Game streaming • Video monitoring / surveillance Slide 3 35pt Internet Protocol Television (IPTV) Font to be used by customers and partners : • Basic requirements: . Random access to pictures 18pt Random Access Period (RAP) should be kept small enough (approximately, 1-15 seconds); Font to be used by customers . Temporal (frame-rate) scalability; and partners : . Error robustness • Optional requirements: . resolution and quality (SNR) scalability Slide 4 35pt Internet Protocol Television (IPTV) Font to be used by customers and partners : Resolution Frame-rate, fps Picture access mode 2160p (4K),3840x2160 60 RA 18pt 1080p, 1920x1080 24, 50, 60 RA 1080i, 1920x1080 30 (60 fields per second) RA Font to be used by customers and partners : 720p, 1280x720 50, 60 RA 576p (EDTV), 720x576 25, 50 RA 576i (SDTV), 720x576 25, 30 RA 480p (EDTV), 720x480 50, 60 RA 480i (SDTV), 720x480 25, 30 RA Slide 5 35pt Video conferencing Font to be used by customers and partners : • Basic requirements: . Delay should be kept as low as possible 18pt The preferable and maximum delay values should be less than 100 ms and 350 ms, respectively Font to be used by customers .
    [Show full text]
  • Lecture 11 : Discrete Cosine Transform Moving Into the Frequency Domain
    Lecture 11 : Discrete Cosine Transform Moving into the Frequency Domain Frequency domains can be obtained through the transformation from one (time or spatial) domain to the other (frequency) via Fourier Transform (FT) (see Lecture 3) — MPEG Audio. Discrete Cosine Transform (DCT) (new ) — Heart of JPEG and MPEG Video, MPEG Audio. Note : We mention some image (and video) examples in this section with DCT (in particular) but also the FT is commonly applied to filter multimedia data. External Link: MIT OCW 8.03 Lecture 11 Fourier Analysis Video Recap: Fourier Transform The tool which converts a spatial (real space) description of audio/image data into one in terms of its frequency components is called the Fourier transform. The new version is usually referred to as the Fourier space description of the data. We then essentially process the data: E.g . for filtering basically this means attenuating or setting certain frequencies to zero We then need to convert data back to real audio/imagery to use in our applications. The corresponding inverse transformation which turns a Fourier space description back into a real space one is called the inverse Fourier transform. What do Frequencies Mean in an Image? Large values at high frequency components mean the data is changing rapidly on a short distance scale. E.g .: a page of small font text, brick wall, vegetation. Large low frequency components then the large scale features of the picture are more important. E.g . a single fairly simple object which occupies most of the image. The Road to Compression How do we achieve compression? Low pass filter — ignore high frequency noise components Only store lower frequency components High pass filter — spot gradual changes If changes are too low/slow — eye does not respond so ignore? Low Pass Image Compression Example MATLAB demo, dctdemo.m, (uses DCT) to Load an image Low pass filter in frequency (DCT) space Tune compression via a single slider value n to select coefficients Inverse DCT, subtract input and filtered image to see compression artefacts.
    [Show full text]
  • (A/V Codecs) REDCODE RAW (.R3D) ARRIRAW
    What is a Codec? Codec is a portmanteau of either "Compressor-Decompressor" or "Coder-Decoder," which describes a device or program capable of performing transformations on a data stream or signal. Codecs encode a stream or signal for transmission, storage or encryption and decode it for viewing or editing. Codecs are often used in videoconferencing and streaming media solutions. A video codec converts analog video signals from a video camera into digital signals for transmission. It then converts the digital signals back to analog for display. An audio codec converts analog audio signals from a microphone into digital signals for transmission. It then converts the digital signals back to analog for playing. The raw encoded form of audio and video data is often called essence, to distinguish it from the metadata information that together make up the information content of the stream and any "wrapper" data that is then added to aid access to or improve the robustness of the stream. Most codecs are lossy, in order to get a reasonably small file size. There are lossless codecs as well, but for most purposes the almost imperceptible increase in quality is not worth the considerable increase in data size. The main exception is if the data will undergo more processing in the future, in which case the repeated lossy encoding would damage the eventual quality too much. Many multimedia data streams need to contain both audio and video data, and often some form of metadata that permits synchronization of the audio and video. Each of these three streams may be handled by different programs, processes, or hardware; but for the multimedia data stream to be useful in stored or transmitted form, they must be encapsulated together in a container format.
    [Show full text]
  • HERO6 Black Manual
    USER MANUAL 1 JOIN THE GOPRO MOVEMENT facebook.com/GoPro youtube.com/GoPro twitter.com/GoPro instagram.com/GoPro TABLE OF CONTENTS TABLE OF CONTENTS Your HERO6 Black 6 Time Lapse Mode: Settings 65 Getting Started 8 Time Lapse Mode: Advanced Settings 69 Navigating Your GoPro 17 Advanced Controls 70 Map of Modes and Settings 22 Connecting to an Audio Accessory 80 Capturing Video and Photos 24 Customizing Your GoPro 81 Settings for Your Activities 26 Important Messages 85 QuikCapture 28 Resetting Your Camera 86 Controlling Your GoPro with Your Voice 30 Mounting 87 Playing Back Your Content 34 Removing the Side Door 5 Using Your Camera with an HDTV 37 Maintenance 93 Connecting to Other Devices 39 Battery Information 94 Offloading Your Content 41 Troubleshooting 97 Video Mode: Capture Modes 45 Customer Support 99 Video Mode: Settings 47 Trademarks 99 Video Mode: Advanced Settings 55 HEVC Advance Notice 100 Photo Mode: Capture Modes 57 Regulatory Information 100 Photo Mode: Settings 59 Photo Mode: Advanced Settings 61 Time Lapse Mode: Capture Modes 63 YOUR HERO6 BLACK YOUR HERO6 BLACK 1 2 4 4 3 11 2 12 5 9 6 13 7 8 4 10 4 14 6 1. Shutter Button [ ] 6. Latch Release Button 10. Speaker 2. Camera Status Light 7. USB-C Port 11. Mode Button [ ] 3. Camera Status Screen 8. Micro HDMI Port 12. Battery 4. Microphone (cable not included) 13. microSD Card Slot 5. Side Door 9. Touch Display 14. Battery Door For information about mounting items that are included in the box, see Mounting (page 87).
    [Show full text]
  • The Evolutionof Premium Vascular Ultrasound
    Ultrasound EPIQ 5 The evolution of premium vascular ultrasound Philips EPIQ 5 ultrasound system The new challenges in global healthcare Unprecedented advances in premium ultrasound performance can help address the strains on overburdened hospitals and healthcare systems, which are continually being challenged to provide a higher quality of care cost-effectively. The goal is quick and accurate diagnosis the first time and in less time. Premium ultrasound users today demand improved clinical information from each scan, faster and more consistent exams that are easier to perform, and allow for a high level of confidence, even for technically difficult patients. 2 Performance More confidence in your diagnoses even for your most difficult cases EPIQ 5 is the new direction for premium vascular ultrasound, featuring an exceptional level of clinical performance to meet the challenges of today’s most demanding practices. Our most powerful architecture ever applied to vascular ultrasound EPIQ performance touches all aspects of acoustic acquisition and processing, allowing you to truly experience the evolution to a more definitive modality. Carotid artery bulb Superficial varicose veins 3 The evolution in premium vascular ultrasound Supported by our family of proprietary PureWave transducers and our leading-edge Anatomical Intelligence, this platform offers our highest level of premium performance. Key trends in global ultrasound • The need for more definitive premium • A demand to automate most operator ultrasound with exceptional image functions
    [Show full text]
  • VIDEO Blu-Ray™ Disc Player BP330
    VIDEO Blu-ray™ Disc Player BP330 Internet access lets you stream instant content from Make the most of your HDTV. Blu-ray disc playback Less clutter. More possibilities. Cut loose from Netflix, CinemaNow, Vudu and YouTube direct to delivers exceptional Full HD 1080p video messy wires. Integrated Wi-Fi® connectivity allows your TV — no computer required. performance, along with Bonus-view for a picture-in- you take advantage of Internet access from any picture. available Wi-Fi® connection in its range. VIDEO Blu-ray™ Disc Player BP330 PROFILE & PLAYABLE DISC PLAYABLE AUDIO FORMATS BD Profile 2.0 LPCM Yes USB Playback Yes Dolby® Digital Yes External HDD Playback Yes (via USB) Dolby® Digital Plus Yes BD-ROM/BD-R/BD-RE Yes Dolby® TrueHD Yes DVD-ROM/DVD±R/DVD±RW Yes DTS Yes Audio CD/CD-R/CD-RW Yes DTS-HD Master Audio Yes DTS-CD Yes MPEG 1/2 L2 Yes MP3 Yes LG SMART TV WMA Yes Premium Content Yes AAC Yes Netflix® Yes FLAC Yes YouTube® Yes Amazon® Yes PLAYABLE PHOTO FORMATS Hulu Plus® Yes JPEG Yes Vudu® Yes GIF/Animated GIF Yes CinemaNow® Yes PNG Yes Pandora® Yes MPO Yes Picasa® Yes AccuWeather® Yes CONVENIENCE SIMPLINK™ Yes VIDEO FEATURES Loading Time >10 Sec 1080p Up-scaling Yes LG Remote App Yes (Free download on Google Play and Apple App Store) Noise Reduction Yes Last Scene Memory Yes Deep Color Yes Screen Saver Yes NvYCC Yes Auto Power Off Yes Video Enhancement Yes Parental Lock Yes Yes Yes CONNECTIVITY Wired LAN Yes AUDIO FEATURES Wi-Fi® Built-in Yes Dolby Digital® Down Mix Yes DLNA Certified® Yes Re-Encoder Yes (DTS only) LPCM Conversion
    [Show full text]
  • A Deblocking Filter Hardware Architecture for the High Efficiency
    A Deblocking Filter Hardware Architecture for the High Efficiency Video Coding Standard Cláudio Machado Diniz1, Muhammad Shafique2, Felipe Vogel Dalcin1, Sergio Bampi1, Jörg Henkel2 1Informatics Institute, PPGC, Federal University of Rio Grande do Sul (UFRGS), Porto Alegre, Brazil 2Chair for Embedded Systems (CES), Karlsruhe Institute of Technology (KIT), Germany {cmdiniz, fvdalcin, bampi}@inf.ufrgs.br; {muhammad.shafique, henkel}@kit.edu Abstract—The new deblocking filter (DF) tool of the next encoder configuration: (i) Random Access (RA) configuration1 generation High Efficiency Video Coding (HEVC) standard is with Group of Pictures (GOP) equal to 8 (ii) Intra period2 for one of the most time consuming algorithms in video decoding. In each video sequence is defined as in [8] depending upon the order to achieve real-time performance at low-power specific frame rate of the video sequence, e.g. 24, 30, 50 or 60 consumption, we developed a hardware accelerator for this filter. frames per second (fps); (iii) each sequence is encoded with This paper proposes a high throughput hardware architecture four different Quantization Parameter (QP) values for HEVC deblocking filter employing hardware reuse to QP={22,27,32,37} as defined in the HEVC Common Test accelerate filtering decision units with a low area cost. Our Conditions [8]. Fig. 1 shows the accumulated execution time architecture achieves either higher or equivalent throughput (in % of total decoding time) of all functions included in C++ (4096x2048 @ 60 fps) with 5X-6X lower area compared to state- class TComLoopFilter that implement the DF in HEVC of-the-art deblocking filter architectures. decoder software. DF contributes to up to 5%-18% to the total Keywords—HEVC coding; Deblocking Filter; Hardware decoding time, depending on video sequence and QP.
    [Show full text]
  • Frequently Asked Questions Dolby Digital Plus
    Frequently Asked Questions Dolby® Digital Plus redefines the home theater surround experience for new formats like high-definition video discs. What is Dolby Digital Plus? Dolby® Digital Plus is Dolby’s new-generation multichannel audio technology developed to enhance the premium experience of high-definition media. Built on industry-standard Dolby Digital technology, Dolby Digital Plus as implemented in Blu-ray Disc™ features more channels, less compression, and higher data rates for a warmer, richer, more compelling audio experience than you get from standard-definition DVDs. What other applications are there for Dolby Digital Plus? The advanced spectral coding efficiencies of Dolby Digital Plus enable content producers to deliver high-resolution multichannel soundtracks at lower bit rates than with Dolby Digital. This makes it ideal for emerging bandwidth-critical applications including cable, IPTV, IP streaming, and satellite (DBS) and terrestrial broadcast. Dolby Digital Plus is also a preferred medium for delivering BonusView (Profile 1.1) and BD-Live™ (Profile 2.0) interactive audio content on Blu-ray Disc. Delivering higher quality and more channels at higher bit rates, plus greater efficiency at lower bit rates, Dolby Digital Plus has the flexibility to fulfill the needs of new content delivery formats for years to come. Is Dolby Digital Plus content backward-compatible? Because Dolby Digital Plus is built on core Dolby Digital technologies, content that is encoded with Dolby Digital Plus is fully compatible with the millions of existing home theaters and playback systems worldwide equipped for Dolby Digital playback. Dolby Digital Plus soundtracks are easily converted to a 640 kbps Dolby Digital signal without decoding and reencoding, for output via S/PDIF.
    [Show full text]
  • Contrast-Enhanced High-Frame-Rate Ultrasound Imaging of Flow Patterns in Cardiac Chambers and Deep Vessels
    Delft University of Technology Contrast-Enhanced High-Frame-Rate Ultrasound Imaging of Flow Patterns in Cardiac Chambers and Deep Vessels Vos, Hendrik J.; Voorneveld, Jason D.; Groot Jebbink, Erik; Leow, Chee Hau; Nie, Luzhen; van den Bosch, Annemien E.; Tang, Meng Xing; Freear, Steven; Bosch, Johan G. DOI 10.1016/j.ultrasmedbio.2020.07.022 Publication date 2020 Document Version Final published version Published in Ultrasound in Medicine and Biology Citation (APA) Vos, H. J., Voorneveld, J. D., Groot Jebbink, E., Leow, C. H., Nie, L., van den Bosch, A. E., Tang, M. X., Freear, S., & Bosch, J. G. (2020). Contrast-Enhanced High-Frame-Rate Ultrasound Imaging of Flow Patterns in Cardiac Chambers and Deep Vessels. Ultrasound in Medicine and Biology, 46(11), 2875-2890. https://doi.org/10.1016/j.ultrasmedbio.2020.07.022 Important note To cite this publication, please use the final published version (if applicable). Please check the document version above. Copyright Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons. Takedown policy Please contact us and provide details if you believe this document breaches copyrights. We will remove access to the work immediately and investigate your claim. This work is downloaded from Delft University of Technology. For technical reasons the number of authors shown on this cover page is limited to a maximum of 10. ARTICLE IN PRESS Ultrasound in Med.
    [Show full text]
  • Video Coding Standards
    Module 8 Video Coding Standards Version 2 ECE IIT, Kharagpur Lesson 23 MPEG-1 standards Version 2 ECE IIT, Kharagpur Lesson objectives At the end of this lesson, the students should be able to : 1. Enlist the major video coding standards 2. State the basic objectives of MPEG-1 standard. 3. Enlist the set of constrained parameters in MPEG-1 4. Define the I- P- and B-pictures 5. Present the hierarchical data structure of MPEG-1 6. Define the macroblock modes supported by MPEG-1 23.0 Introduction In lesson 21 and lesson 22, we studied how to perform motion estimation and thereby temporally predict the video frames to exploit significant temporal redundancies present in the video sequence. The error in temporal prediction is encoded by standard transform domain techniques like the DCT, followed by quantization and entropy coding to exploit the spatial and statistical redundancies and achieve significant video compression. The video codecs therefore follow a hybrid coding structure in which DPCM is adopted in temporal domain and DCT or other transform domain techniques in spatial domain. Efforts to standardize video data exchange via storage media or via communication networks are actively in progress since early 1980s. A number of international video and audio standardization activities started within the International Telephone Consultative Committee (CCITT), followed by the International Radio Consultative Committee (CCIR), and the International Standards Organization / International Electrotechnical Commission (ISO/IEC). An experts group, known as the Motion Pictures Expects Group (MPEG) was established in 1988 in the framework of the Joint ISO/IEC Technical Committee with an objective to develop standards for coded representation of moving pictures, associated audio, and their combination for storage and retrieval of digital media.
    [Show full text]