4. (A) What Is the Distinction Between Lossy and Lossless Data Compression?
Total Page:16
File Type:pdf, Size:1020Kb
?a) What is the distinction between lossy and lossless data compression) .4 Lossless preserves data undergoing compression, Lossy compression aims to obtain the best possible fidelity for a given bit-rate or minimizing the bit-rate to achieve a given fidelity measure but will not produce a complete facsimile of the .original data
Briefly describe the four basic types of data redundancy that data compression _10 .algorithms can apply to audio, image and video signals :Types of Compression 4 .Temporal -- in 1D data, 1D signals, Audio etc Spatial -- correlation between neighbouring pixels or data items Spectral -- correlation between colour or luminescence components. This uses the frequency domain to exploit relationships between frequency of change in .data Psycho-visual, psycho-acoustic -- exploit perceptual properties of the human .visual system or aural system to compress data
?a) What is the distinction between lossy and lossless data compression) .3 Lossless Compression Where data is compressed and can be reconstituted (uncompressed) without loss of detail or information. These are referred to as bit-preserving or reversible .compression systems also Lossy Compression where the aim is to obtain the best possible fidelity for a given bit-rate or minimizing the bit-rate to achieve a given fidelity measure. Video and audio .compression techniques are most suited to this form of compression
a) What is meant by the terms Multimedia and Hypermedia? Distinguish) .1 .between these two concepts Multimedia ---- An Application which uses a collection of multiple media sources .e.g. text, graphics, images, sound/audio, animation and/or video Hypermedia --- An application which uses associative relationships among information contained within multiple media data for the purpose of facilitating .access to, and manipulation of, the information encapsulated by the data
b) What is meant by the terms static media and dynamic media? Give two) .examples of each type of media Static Media – does not change over time, e.g. text, graphics .Dynamic Media --- Time dependent (Temporal), e.g. Video, sound, animation ?a) Why is file or data compression necessary for Multimedia activities) .2 Multimedia files are very large therefore for storage, file transfer etc. file sizes need to be reduced often as part of the file format. Text and other files may also be .encoded/compressed for email and other applications
?a) What is MIDI? How is a basic MIDI message structured) _17 MIDI: a protocol that enables computer, synthesizers, keyboards, and other .musical or (even) multimedia devices to communicate with each other :MIDI MESSAGE .MIDI message includes a status byte and up to two data bytes Status byte .The most significant bit of status byte is set to 1 The 4 low-order bits identify which channel it belongs to (four bits produce 16 .(possible channels .The 3 remaining bits identify the message .The most significant bit of data byte is set to 0
In what ways can MIDI be used effectively in Multimedia Applications, as _18 ?opposed to strictly musical applications :Many Application Low Bandwidth/(Low Quality?) Music on Web, Quicktime etc supports Midi musical instrument set Sound Effectts --- Low Bandwidth alternative to audio samples, Sound Set part of GM soundset Control of external devices --- e.g Synchronistaion of Video and Audio SMPTE), Midi System Exclusive, AUDIO RECORDERS, SAMPLERS) Control of synthesis --- envelope control etc (MPEG 4 Compression control --- see Part (c Digital Audio
?How can MIDI be used with modern data compression techniques _19 ?Briefly describe how such compression techniques may be implemented We have seen the need for compression already in Digital Audio -- Large Data Files Basic Ideas of compression (see next Chapter) used as integral part of audio -- format .MP3, real audio etc Mpeg-4 audio -- actually combines compression synthesis and midi to have a massive .impact on compression Midi, Synthesis encode what note to play and how to play it with a small number of .parameters -- Much greater reduction than simply having some encoded bits of audio .Responsibility to create audio delegated to generation side :MPEG-4 comprises of 6 Structured Audio tools are SAOL the Structured Audio Orchestra Language SASL the Structured Audio Score Language SASBF the Structured Audio Sample Bank Format a set of MIDI semantics which describes how to control SAOL with MIDI a scheduler which describes how to take the above parts and create sound the AudioBIFS part of BIFS, which lets you make audio soundtracks in MPEG-4 using a variety of tools and effects-processing techniques :MIDI IS the control language for the synthesis part As well as controlling synthesis with SASL scripts, it can be controlled with MIDI files and scores in MPEG-4. MIDI is today's most commonly used representation for music score data, and many sophisticated authoring tools (such as sequencers) work .with MIDI The MIDI syntax is external to the MPEG-4 Structured Audio standard; only references to the MIDI Manufacturers Association's definition in the standard. But in order to make the MIDI controls work right in the MPEG context, some semantics what the instructions "mean") have been redefined in MPEG-4. The new semantics) .are carefully defined as part of the MPEG-4 specification
.a) Give a definition of multimedia and a multimedia system) .1 Multimedia is the field concerned with the computer-controlled integration of text, graphics, drawings, still and moving images (Video), animation, audio, and ,any other media where every type of information can be represented, stored .transmitted and processed digitally A Multimedia System is a system capable of processing multimedia data and .applications
b) What are the key distinctions between multimedia data and more conventional ) ?types of media ,Multimedia systems deal with the generation, manipulation, storage, presentation .and communication of information in digital form .The data may be in a variety of formats: text, graphics, images, audio, video - A majority of this data is large and the different media may need synchronisation .the data may have temporal relationships as an integral property - ,Some media is time independent or static or discrete media: normal data, text .single images, graphics are examples Video, animation and audio are examples of continuous media
What key issues or problems does a multimedia system have to deal with when _13 ?handling multimedia data :A Multimedia system has four basic characteristics .Multimedia systems must be computer controlled .Multimedia systems are integrated .The information they handle must be represented digitally .The interface to the final presentation of media is usually interactive CM0340 3 Multimedia systems may have to render a variety of media at the same instant -- a distinction from normal applications. There is a temporal relationship between many forms of media (e.g. Video and Audio. There 2 are forms of problems here Sequencing within the media -- playing frames in correct order/time frame in video Synchronisation -- inter-media scheduling (e.g. Video and Audio). Lip synchronisation is clearly important for humans to watch playback of video (and audio and even animation and audio. Ever tried watching an out of (lip ?sync film for a long time :The key issues multimedia systems need to deal with here are .How to represent and store temporal information How to strictly maintain the temporal relationships on play back/retrieval .What process are involved in the above -- Data has to represented digitally so many initial source of data needs to be digitise translated from analog source to digital representation. The will involve scanning graphics, still images), sampling (audio/video) although digital cameras now exist for) .direct scene to digital capture of images and video The data is large several Mb easily for audio and video -- therefore storage, transfer bandwidth) and processing overheads are high. Data compression techniques very) .common
List three distinct models of colour used in Multimedia. Explain why there _9 .are a number of different colour models exploited in multimedia data formats :Possible models RGB CIE Chromaticity YIQ Colour Space (YUV (YCrCb CMY/CMYK Different models reflect need to represent colour in a perceptually relevant model .for effective compression Different models also due to evolution of colour from Video (YIQ,YUV), Display .RGB) and Print (CMYK) media requirements)
.Give one example of a compression algorithm for each class :EXAMPLES ,Temporal — Any Audio/Video compression method, Zero length suppression ,pattern substitution, Pulse Code Modulation (a few variants), MPEG Audio [MPEG Video, H.264 [1 .Spatial —Any Image/Video compression algorithm, GIF, JPEG, MPEG, H.264 [1] [Spectral — JPEG, MPEG, H.264. [1 [Psycho-visual — MPEG audio, MPEG Video, JPEG (colour conversion). [1 b) What advantage does arithmetic coding offer over Huffman coding for data) ?compression Huffman coding assumes an integer number (k) of bits for each symbol hence k [is never less than 1 [1 Arithmetic coding can represent fractional number of bits and can achieve better [compression ratios. [1
.Briefly state an algorithm for arithmetic decoding _2 :SEEN In LECTURE Coding The idea behind arithmetic coding is To have a probability line, 0–1, and _ ,Assign to every symbol a range in this line based on its probability _ 2 CM0340 SOLNS .Order in terms of probability highest first _ .Note: The higher the probability, the higher range which assigns to it _ For each symbol in the sequence assign a code based in symbols probability and :then subdivide for all the symbols ;range = high - low ;high = low + range * high_range of the symbol being coded ;low = low + range * low_range of the symbol being coded (Decoding is the opposite so need to work out (unseen in lectures :For current code value [look up in table and assign symbol [1 _ Eliminate symbol effect by subtracting the low value in the range and divide _ [by range [2 [Repeat above two steps until zero reached — see last part of problem. [2 _ Total 5 marks—Unseen. (Coding algorithm discussed in lectures, decoding (simply mentioned as the reverse process
c) Why are I-frames inserted into the compressed output stream relatively) ?frequently Differences between frames get too large — large errors – hard to track fast .blocks etc. So need to restart card with a new I-frame
?What is the key difference between I-Frames, P-Frames and B-Frames _3 I-Frame — Basic Reference frame for each Group of pictures — essentially a [JPEG Compressed image. [1 [P-Frame — Coded forward Difference frame w.r.t last I or P frame [1 [B-Frame — Coded backward Difference frame w.r.t last I or P frame [1
?Q3. (a) What is the difference between reverb and echo [Echo — implies a distinct, delayed version of a sound, [1 Reverb — each delayed sound wave arrives in such a short period of time such [that we do not perceive each reflection as a copy of the original sound. [1
Q3. (a) What are critical bands in relation to the human ear’s perception of ?sound Frequncy masking occurs in human ear where certain frequencies are masked by neighbouring frequencies. Range of closeness for frequency masking depends on the frequencies and relative amplitudes. Each band where frequencies are masked is called the Critical Band
?b) How does MPEG audio compression achieve critical band approximation) [Create Critical Band Filter Banks [1 [Fourier Transform Audio Data [1 [Bandpass filter data according to know Critical Bandwidths[1
Briefly explain how the human visual system senses colour. How is _8 colour exploited in the compression of multimedia graphics, images and ?video ”The eye is basically just a “biological camera (Eye through lens etc focused light onto the Retina (back of eye Retina consists of neurons that fire nerve signals proportional to incident light .Each neuron is either a rod or a cone. Rods are not sensitive to colour Cones organized in banks that sense red green and blue :Multimedia Context Since Rods do not sense colour only sense luminosity intensity Eye is more sensitve to luminence than colour. Also Eye is more sensitive to red and green and blue (this is due to evolution need to see fellow humans where blue is not (prevalent in skin hues So any Multimedia compression techniques should use colour representation that presents colour in a way that models Human Visual system. We can then encode luminence is high bandwith (more bits) than colour as this is much more .perceptually relevant