Symbols A-Law, 148, 152, 166 Compander, 232 Μ-Law, 148, 152, 166

Total Page:16

File Type:pdf, Size:1020Kb

Symbols A-Law, 148, 152, 166 Compander, 232 Μ-Law, 148, 152, 166 Index Symbols Lingo script, 46 A-law, 148, 152, 166 object, 47 compander, 232 tweening, 44 µ-law, 148, 152, 166, 435 Adobe Flash, 47 compander, 232 animation, 50 2D mesh, 375 symbol, 49 geometry coding, 376 window, 48 motion coding, 378 Adobe Flash Video, 619 object coding, 375 Adobe Photoshop, 41 2D object animation, 379 alpha channel, 41 3D model-based coding, 381 magic wand tool, 42 3D percept, 130 Adobe Premiere, 39 3D polygon mesh, 383 timeline window, 39 3D video and TV, 130 ADPCM (adaptive differential pulse code 3G, 582, 583, 589 modulation), 165, 175, 435–437 G3G (Global 3G), 583 Affine transform, 380 4G, 584, 589 Alias, 142–144 Amazon web services (AWS), 14, 649, 657, A 659 AC (Alternate Current), 236 Amazon EBS, 652 Access network, 486, 489, 494, 564 Amazon EC2, 647, 650, 660 Access point (AP), 576, 586–588, 608 Amazon machine image (AMI), 651 Access time, 545 Amazon S3, 645, 647, 649, 660 Active pixel, 118 AWS region, 649, 651 Active video line, 118 Cloudfront, 657 AD (analog-to-digital) converter, 149, 175 AMPS (advanced mobile phone system), 577 Adaptive compression algorithms, 196 Anaglyph 3D, 132 Adaptive Huffman coding, 196 Analog display interface, 126 Adobe Director, 42 Analog video, 115 3D Sprite, 47 Animation, 18 animation, 44 Autodesk 3ds Max, 19 control, 44 Autodesk Maya, 19 Imaging Lingo, 43 Autodesk Softimage, 19 Lingo, 43 DirectX, 18 Z.-N. Li et al., Fundamentals of Multimedia, 715 Texts in Computer Science, DOI: 10.1007/978-3-319-05290-8, © Springer International Publishing Switzerland 2014 716 Index Java3D, 18 Bi-level image compression standards, also see OpenGL, 18 JBIG, 309 Anti-aliasing, 143 Bilinear interpolation, 346, 355 filter, 143 Binary tree, 190 APC (adaptive predictive coding), 176 Bitmap, 58 Application-layer multicast, 506, 553 Bitplane, 59 end-system multicast (ESM), 555 Bitrate, 151, 195, 225, 267, 270, 317, 325, 332, multi-tree overlay, 556 430, 507, 509, 551, 590 single tree overlay, 555, 556 Block codes, 594 Arithmetic coding, 205, 372 Block-based coding, 359, 363 adaptive arithmetic coding, 215, 216 Blu-ray, 14 basic algorithm, 206 Bluetooth, 589 binary arithmetic coding, 214, 413, 415 BMP (bitmap), 75 integer implementation, 214 Broadcast, 503, 550 scaling and incremental coding, 210 Buffer management, 514 tag, 208, 210 peak bitrate, 515 Aspect ratio, 118, 123, 125 prefetch buffer, 515 ATM (asynchronous transfer mode), 496, 511 ATV (advanced TV), 348 C Audio compression standard C-BIRD, 680 G.711, 437, 523 search by color, 680, 682 G.721, 436 search by illumination invariance, 685 G.722, 523 search by object model, 686 G.723, 436 search by texture, 683 G.723.1, 447, 449 CABAC (context-adaptive binary arithmetic G.726, 436, 437 coding), 396, 419 G.727, 436 Cable modem, 486, 492 G.728, 436, 450 Cable TV network, 489, 492 G.729, 436, 447, 450 CAE (context-based arithmetic encoding), 370 Audio filtering, 150 Camera system, 85 CAVLC (context-adaptive variable length Autostereoscopic display device, 135 coding), 396 lenticular lens, 135 CBIR (content-based image retrieval), 675, parallax barrier, 135 680, 703 AVC (advanced video coding, also see 3D shapes and objects, 677 H.264/AVC), 395 early CBIR systems, 678 AVCHD (advanced video coding high histogram intersection, 678 definition), 18 human activity, 700, 703 AVI (audio video interleave), 39 key technologies, 692 quality-aware, 706 B quantifying search results, 688 BAB (binary alpha block), 370 video retrieval, 679, 697 Bag of words (BoW), 694 CCIR (consultative committee for international Band-limited signal, 143 radio), 122 Band-limiting filter, 151 CCITT (international telegraph and telephone Band-pass filter, 151, 158, 166 consultative committee), 168, 325 Bandwidth, 151, 166, 489, 512, 531, 566, 577, CDMA (code division multiple access), 578, 618, 628, 645, 652, 662, 670 609 Base station (BS), 576, 579, 585, 608 cdma2000, 583 BCH (Bose-Chaudhuri-Hocquenghem) codes, WCDMA (wideband CDMA), 582, 609 595 CELP (code excited linear prediction), 444 Index 717 adaptive codebook, 445 spectra sensitivity of the eye, 83 LSF (line spectrum frequency), 448 visible light, 82 LSP (line spectrum pair), 447 Color-matching function, 88, 89 LTP (long time prediction), 445 Commission internationale de L’eclairage stochastic codebook, 448 (CIE), 89 STP (short-time prediction), 445 Compression Checksum, 498, 500, 591 lossless, 62, 74, 186 Chroma subsampling, 122, 407, 417, 419 lossy, 62, 74, 186, 282 Chromaticity, 90 ratio, 74, 186 diagram, 89 speech, 166 Chrominance, 106 Compressor function, 231 CIF (common intermediate format), 123 Cones, 83 Circuit switching, 495 Constant bit rate (CBR), 501, 515, 546 Client/server, 531, 532, 540, 567 Content delivery network (CDN), also see Cloud computing, 645 content distribution network (CDN), cloud gaming, 667 539 computation offloading, 661 Content distribution network (CDN), 506, 539, infrastructure as a service (IaaS), 647 541, 542, 565, 629, 657 platform as a service (PaaS), 648 Akamai, 506, 542, 660 private cloud, 647 Context modeling, 305 public cloud, 647 Continuous Fourier Transform (CFT), 247 software as a service(SaaS), 648 Continuous Wavelet Transform (CWT), 252, Cloud gaming, 665 256 Gaikai, 15, 668 Convolutional codes, 596 Onlive, 668 Coordinated live streaming and storage sharing Clustering, 67 (COOLS), 638 CMY, 102 CPU (central processing unit), 651, 652, 661, CMYK, 103 662, 665 Codec, 186, 477 CRC (cyclic redundancy check), 488, 592 Coder mapping, 165 CRT (cathode ray tube), 86, 116 Codeword, 186 CSS (cascading style sheets), 11 Coding efficiency, 430 Color D camera-dependent, 100 DA (digital-to-analog) converter, 149 cycling, 65 Datagram, 498 density, 682 DB (decibel), 144 histogram, 64, 680 DC (Direct Current), 236 HSV, 101 Decoder mapping, 165 layout, 682 Deinterlacing, 116 lookup table (LUT), 67 Delaunay mesh, 377 monitor specification, 93 Dictionary-based coding, 200 multi-ink printers, 104 Difference of Gaussian (DOG), 693 multisensor cameras, 100 Differential coding, 168, 218 palette, 65 Differentiated service (DiffServ), 513 palette animation, 65 diffServ code (DS), 513, 514 picker, 65 per-hop behavior (PHB), 513, 514 primaries, 88 Digital audio, 16 sRGB, 101 Adobe Audition, 16 subcarrier, 119 coding of, 165 Color science, 81 quantization and transmission, 164 light and spectra, 81 Sound Forge, 16 718 Index Digital display interface, 128 E Digital library, 185 EBCOT (Embedded block coding with Digital subscriber line (DSL), 491 optimized truncation), 270, 295, ADSL (asymmetrical DSL), 486, 491, 493, 298, 301, 303 564 EDTV (enhanced definition TV), 126 Digitization of sound, 139 End-to-end argument, 503 Discrete Cosine Transform (DCT), 234, 236, Entropy, 186, 187 241 coding, 187, 189 1D, 235, 236 Epidemic model, 634 2D, 234, 235, 281 Error concealment, 500, 603 2D basis function, 244 Error detection, 590 2D matrix implementation, 245 Error-resilient coding, 597 2D separable basis, 244 Error-resilient entropy coding (EREC), 600 basis function, 236, 242 Ethernet, 488 comparison to DFT, 247 CSMA/CD (carrier sense multiple access DCT-matrix, 246, 400, 424 with collision detection), 489 Discrete Fourier Transform (DFT), 247 Euler’s formula, 247 Discrete Sine Transform (DST), 419, 425 Excited purity, 93 Discrete Wavelet Transform (DWT), 252, 259 EXIF (exchangeable image file), 76 Disparity, 131 Exp-Golomb code, 215, 409, 410, 413 gradient, 136 Expander function, 232 manipulation, 136 mapping, 136 Extended Huffman coding, 194 range, 136 EZW (Embedded Zerotree Wavelet), 270–273 sensitivity, 136 velocity, 136 F Dispersion, 82 F-score, 711 Distortion measure, 225 Facebook, 14, 617, 619 MSE (Mean Square Error), 226 Fax standards PSNR (Peak Signal-to-Noise Ratio), 226 G3, 309 SNR (Signal-to-Noise Ratio), 226 G4, 309 Dithering, 59, 66 FDMA (frequency division multiple access), dither matrix, 60 577, 578, 580 ordered dither, 61 Firewall, 486, 502, 503 DM (delta modulation), 174 First person shooter (FPS) game, 666 adaptive, 175 FM (frequency modulation), 152, 154 uniform, 175 Forward error correction (FEC), 593 DMT (discrete multi-tone), 491 Fourier transform, 251 Domain name system (DNS), 496 Frame buffer, 59, 100 DPCM (differential pulse code modulation), 165, 168, 171, 289 Frame-based coding, 359, 363 Dropbox, 645 Free rider, 564 DV video (digital video), 122 Frequency, 236 DVB (digital video broadcasting), 544 frequency response, 236, 244 DVB-MHP (multimedia home platform), frequency spectrum, 236 544 spatial frequency, 236, 244 DVD, 9, 348, 349, 475, 479, 658 Frequency hopping (FH), 589 Dynamic adaptive streaming over HTTP FTTH (fiber-to-the-home), 493 (DASH), 480, 565, 660 FTTN (fiber-to-the-neighborhood), 493 media presentation description (MPD), 565 FTTN (fiber-to-the-node), 493 Dynamic range, 168 Fundamental frequency, 140 Index 719 G H.263+, 336 Gamma correction, 86, 96, 100 H.263++, 336 Gamut, 94 H.264, 395, 620, 668 printer, 103 CABAC, 409, 413 Gaussian distribution, 277 CAVLC, 409, 411 Generalized Markup language (GML), 10 entropy coding, 409 GIF (graphics interchange format), 69 group of pictures (GOP), 399 color map, 71 hierarchical prediction structure, 399 GIF87, 72 in-loop deblocking filtering, 407, 408 GIF89, 72 integer transform, 401 interlacing, 72 intra coding, 404 screen descriptor, 69 intra spatial prediction, 404 GIF(graphics interchange format) motion compensation, 396 animation, 19 multiple reference frame, 399 Global motion compensation, 374 MVC (multiview video coding), 417 Gossip algorithm, 558, 560 profiles, 415 GPRS (general packet radio service), 580 baseline profile, 415 GPS
Recommended publications
  • JPEG Image Compression
    JPEG Image Compression JPEG is the standard lossy image compression Based on Discrete Cosine Transform (DCT) Comes from the Joint Photographic Experts Group Formed in1986 and issued format in 1992 Consider an image of 73,242 pixels with 24 bit color Requires 219,726 bytes Image Quality Size Rati o Highest 81,447 2.7: Extremely minor Q=100 1 artifacts High 14,679 15:1 Initial signs of Quality subimage Q=50 artifacts Medium 9,407 23:1 Stronger Q artifacts; loss of high frequency information Low 4,787 46:1 Severe high frequency loss leads to obvious artifacts on subimage boundaries ("macroblocking ") Lowest 1,523 144: Extreme loss of 1 color and detail; the leaves are nearly unrecognizable JPEG How it works Begin with a color translation RGB goes to Y′CBCR Luma and two Chroma colors Y is brightness CB is B-Y CR is R-Y Downsample or Chroma Subsampling Chroma data resolutions reduced by 2 or 3 Eye is less sensitive to fine color details than to brightness Block splitting Each channel broken into 8x8 blocks no subsampling Or 16x8 most common at medium compression Or 16x16 Must fill in remaining areas of incomplete blocks This gives the values DCT - centering Center the data about 0 Range is now -128 to 127 Middle is zero Discrete cosine transform formula Apply as 2D DCT using the formula Creates a new matrix Top left (largest) is the DC coefficient constant component Gives basic hue for the block Remaining 63 are AC coefficients Discrete cosine transform The DCT transforms an 8×8 block of input values to a linear combination of these 64 patterns.
    [Show full text]
  • COLOR SPACE MODELS for VIDEO and CHROMA SUBSAMPLING
    COLOR SPACE MODELS for VIDEO and CHROMA SUBSAMPLING Color space A color model is an abstract mathematical model describing the way colors can be represented as tuples of numbers, typically as three or four values or color components (e.g. RGB and CMYK are color models). However, a color model with no associated mapping function to an absolute color space is a more or less arbitrary color system with little connection to the requirements of any given application. Adding a certain mapping function between the color model and a certain reference color space results in a definite "footprint" within the reference color space. This "footprint" is known as a gamut, and, in combination with the color model, defines a new color space. For example, Adobe RGB and sRGB are two different absolute color spaces, both based on the RGB model. In the most generic sense of the definition above, color spaces can be defined without the use of a color model. These spaces, such as Pantone, are in effect a given set of names or numbers which are defined by the existence of a corresponding set of physical color swatches. This article focuses on the mathematical model concept. Understanding the concept Most people have heard that a wide range of colors can be created by the primary colors red, blue, and yellow, if working with paints. Those colors then define a color space. We can specify the amount of red color as the X axis, the amount of blue as the Y axis, and the amount of yellow as the Z axis, giving us a three-dimensional space, wherein every possible color has a unique position.
    [Show full text]
  • 1. in the New Document Dialog Box, on the General Tab, for Type, Choose Flash Document, Then Click______
    1. In the New Document dialog box, on the General tab, for type, choose Flash Document, then click____________. 1. Tab 2. Ok 3. Delete 4. Save 2. Specify the export ________ for classes in the movie. 1. Frame 2. Class 3. Loading 4. Main 3. To help manage the files in a large application, flash MX professional 2004 supports the concept of _________. 1. Files 2. Projects 3. Flash 4. Player 4. In the AppName directory, create a subdirectory named_______. 1. Source 2. Flash documents 3. Source code 4. Classes 5. In Flash, most applications are _________ and include __________ user interfaces. 1. Visual, Graphical 2. Visual, Flash 3. Graphical, Flash 4. Visual, AppName 6. Test locally by opening ________ in your web browser. 1. AppName.fla 2. AppName.html 3. AppName.swf 4. AppName 7. The AppName directory will contain everything in our project, including _________ and _____________ . 1. Source code, Final input 2. Input, Output 3. Source code, Final output 4. Source code, everything 8. In the AppName directory, create a subdirectory named_______. 1. Compiled application 2. Deploy 3. Final output 4. Source code 9. Every Flash application must include at least one ______________. 1. Flash document 2. AppName 3. Deploy 4. Source 10. In the AppName/Source directory, create a subdirectory named __________. 1. Source 2. Com 3. Some domain 4. AppName 11. In the AppName/Source/Com directory, create a sub directory named ______ 1. Some domain 2. Com 3. AppName 4. Source 12. A project is group of related _________ that can be managed via the project panel in the flash.
    [Show full text]
  • Video Coding Standards
    Video coding standards Video signals represent sequences of images or frames which can be transmitted with a rate from 15 to 60 frames per second (fps), that provides the illusion of motion in the displayed signal. Unlike images they contain the so-called temporal redundancy. Temporal redundancy arises from repeated objects in consecutive frames of the video sequence. Such objects can remain, they can move horizontally, vertically, or any combination of directions (translation movement), they can fade in and out, and they can disappear from the image as they move out of view. Temporal redundancy Motion compensation A motion compensation technique is used to compensate the temporal redundancy of video sequence. The main idea of the this method is to predict the displacement of group of pixels (usually block of pixels) from their position in the previous frame. Information about this displacement is represented by motion vectors which are transmitted together with the DCT coded difference between the predicted and the original images. Motion compensation Image VLC DCT Quantizer Buffer - Dequantizer Coded DCT coefficients. IDCT Motion compensated predictor Motion Motion vectors estimation Decoded VL image Buffer Dequantizer IDCT decoder Motion compensated predictor Motion vectors Motion compensation Previous frame Current frame Set of macroblocks of the previous frame used to predict the selected macroblock of the current frame Motion compensation Previous frame Current frame Macroblocks of previous frame used to predict current frame Motion compensation Each 16x16 pixel macroblock in the current frame is compared with a set of macroblocks in the previous frame to determine the one that best predicts the current macroblock.
    [Show full text]
  • The Video Codecs Behind Modern Remote Display Protocols Rody Kossen
    The video codecs behind modern remote display protocols Rody Kossen 1 Rody Kossen - Consultant [email protected] @r_kossen rody-kossen-186b4b40 www.rodykossen.com 2 Agenda The basics H264 – The magician Other codecs Hardware support (NVIDIA) The basics 4 Let’s start with some terms YCbCr RGB Bit Depth YUV 5 Colors 6 7 8 9 Visible colors 10 Color Bit-Depth = Amount of different colors Used in 2 ways: > The number of bits used to indicate the color of a single pixel, for example 24 bit > Number of bits used for each color component of a single pixel (Red Green Blue), for example 8 bit Can be combined with Alpha, which describes transparency Most commonly used: > High Color – 16 bit = 5 bits per color + 1 unused bit = 32.768 colors = 2^15 > True Color – 24 bit Color + 8 bit Alpha = 8 bits per color = 16.777.216 colors > Deep color – 30 bit Color + 10 bit Alpha = 10 bits per color = 1.073 billion colors 11 Color Bit-Depth 12 Color Gamut Describes the subset of colors which can be displayed in relation to the human eye Standardized gamut: > Rec.709 (Blu-ray) > Rec.2020 (Ultra HD Blu-ray) > Adobe RGB > sRBG 13 Compression Lossy > Irreversible compression > Used to reduce data size > Examples: MP3, JPEG, MPEG-4 Lossless > Reversable compression > Examples: ZIP, PNG, FLAC or Dolby TrueHD Visually Lossless > Irreversible compression > Difference can’t be seen by the Human Eye 14 YCbCr Encoding RGB uses Red Green Blue values to describe a color YCbCr is a different way of storing color: > Y = Luma or Brightness of the Color > Cr = Chroma difference
    [Show full text]
  • CHAPTER 10 Basic Video Compression Techniques Contents
    Multimedia Network Lab. CHAPTER 10 Basic Video Compression Techniques Multimedia Network Lab. Prof. Sang-Jo Yoo http://multinet.inha.ac.kr The Graduate School of Information Technology and Telecommunications, INHA University http://multinet.inha.ac.kr Multimedia Network Lab. Contents 10.1 Introduction to Video Compression 10.2 Video Compression with Motion Compensation 10.3 Search for Motion Vectors 10.4 H.261 10.5 H.263 The Graduate School of Information Technology and Telecommunications, INHA University 2 http://multinet.inha.ac.kr Multimedia Network Lab. 10.1 Introduction to Video Compression A video consists of a time-ordered sequence of frames, i.e.,images. Why we need a video compression CIF (352x288) : 35 Mbps HDTV: 1Gbps An obvious solution to video compression would be predictive coding based on previous frames. Compression proceeds by subtracting images: subtract in time order and code the residual error. It can be done even better by searching for just the right parts of the image to subtract from the previous frame. Motion estimation: looking for the right part of motion Motion compensation: shifting pieces of the frame The Graduate School of Information Technology and Telecommunications, INHA University 3 http://multinet.inha.ac.kr Multimedia Network Lab. 10.2 Video Compression with Motion Compensation Consecutive frames in a video are similar – temporal redundancy Frame rate of the video is relatively high (30 frames/sec) Camera parameters (position, angle) usually do not change rapidly. Temporal redundancy is exploited so that not every frame of the video needs to be coded independently as a new image. The difference between the current frame and other frame(s) in the sequence will be coded – small values and low entropy, good for compression.
    [Show full text]
  • Multiple Reference Motion Compensation: a Tutorial Introduction and Survey Contents
    Foundations and TrendsR in Signal Processing Vol. 2, No. 4 (2008) 247–364 c 2009 A. Leontaris, P. C. Cosman and A. M. Tourapis DOI: 10.1561/2000000019 Multiple Reference Motion Compensation: A Tutorial Introduction and Survey By Athanasios Leontaris, Pamela C. Cosman and Alexis M. Tourapis Contents 1 Introduction 248 1.1 Motion-Compensated Prediction 249 1.2 Outline 254 2 Background, Mosaic, and Library Coding 256 2.1 Background Updating and Replenishment 257 2.2 Mosaics Generated Through Global Motion Models 261 2.3 Composite Memories 264 3 Multiple Reference Frame Motion Compensation 268 3.1 A Brief Historical Perspective 268 3.2 Advantages of Multiple Reference Frames 270 3.3 Multiple Reference Frame Prediction 271 3.4 Multiple Reference Frames in Standards 277 3.5 Interpolation for Motion Compensated Prediction 281 3.6 Weighted Prediction and Multiple References 284 3.7 Scalable and Multiple-View Coding 286 4 Multihypothesis Motion-Compensated Prediction 290 4.1 Bi-Directional Prediction and Generalized Bi-Prediction 291 4.2 Overlapped Block Motion Compensation 294 4.3 Hypothesis Selection Optimization 296 4.4 Multihypothesis Prediction in the Frequency Domain 298 4.5 Theoretical Insight 298 5 Fast Multiple-Frame Motion Estimation Algorithms 301 5.1 Multiresolution and Hierarchical Search 302 5.2 Fast Search using Mathematical Inequalities 303 5.3 Motion Information Re-Use and Motion Composition 304 5.4 Simplex and Constrained Minimization 306 5.5 Zonal and Center-biased Algorithms 307 5.6 Fractional-pixel Texture Shifts or Aliasing
    [Show full text]
  • Overview: Motion-Compensated Coding
    Overview: motion-compensated coding Motion-compensated prediction Motion-compensated hybrid coding Motion estimation by block-matching Motion estimation with sub-pixel accuracy Power spectral density of the motion-compensated prediction error Rate-distortion analysis Loop filter Motion compensated coding with sub-pixel accuracy Rate-constrained motion estimation Bernd Girod: EE398B Image Communication II Motion Compensated Coding no. 1 Motion-compensated prediction previous frame stationary Δt background current frame x time t y moving object ⎛ d x ⎞ „Displacement vector“ ⎜ d ⎟ shifted ⎝ y ⎠ object Prediction for the luminance signal S(x,y,t) within the moving object: ˆ S (x, y,t) = S(x − dx , y − dy ,t − Δt) Bernd Girod: EE398B Image Communication II Motion Compensated Coding no. 2 Combining transform coding and prediction Transform domain prediction Space domain prediction Q Q T - T - T T −1 T −1 T −1 PT T PTS T T −1 T−1 PT PS Bernd Girod: EE398B Image Communication II Motion Compensated Coding no. 3 Motion-compensated hybrid coder Coder Control Control Data Intra-frame DCT DCT Coder - Coefficients Decoder Intra-frame Decoder 0 Motion- Compensated Intra/Inter Predictor Motion Data Motion Estimator Bernd Girod: EE398B Image Communication II Motion Compensated Coding no. 4 Motion-compensated hybrid decoder Control Data DCT Coefficients Decoder Intra-frame Decoder 0 Motion- Compensated Intra/Inter Predictor Motion Data Bernd Girod: EE398B Image Communication II Motion Compensated Coding no. 5 Block-matching algorithm search range in Subdivide current reference frame frame into blocks. Sk −1 Find one displacement vector for each block. Within a search range, find a best „match“ that minimizes an error measure.
    [Show full text]
  • Efficient Video Coding with Motion-Compensated Orthogonal
    Efficient Video Coding with Motion-Compensated Orthogonal Transforms DU LIU Master’s Degree Project Stockholm, Sweden 2011 XR-EE-SIP 2011:011 Efficient Video Coding with Motion-Compensated Orthogonal Transforms Du Liu July, 2011 Abstract Well-known standard hybrid coding techniques utilize the concept of motion- compensated predictive coding in a closed-loop. The resulting coding de- pendencies are a major challenge for packet-based networks like the Internet. On the other hand, subband coding techniques avoid the dependencies of predictive coding and are able to generate video streams that better match packet-based networks. An interesting class for subband coding is the so- called motion-compensated orthogonal transform. It generates orthogonal subband coefficients for arbitrary underlying motion fields. In this project, a theoretical lossless signal model based on Gaussian distribution is proposed. It is possible to obtain the optimal rate allocation from this model. Addition- ally, a rate-distortion efficient video coding scheme is developed that takes advantage of motion-compensated orthogonal transforms. The scheme com- bines multiple types of motion-compensated orthogonal transforms, variable block size, and half-pel accurate motion compensation. The experimental results show that this scheme outperforms individual motion-compensated orthogonal transforms. i Acknowledgements This thesis was carried out at Sound and Image Processing Lab, School of Electrical Engineering, KTH. I would like to express my appreciation to my supervisor Markus Flierl for the opportunity of doing this thesis. I am grateful for his patience and valuable suggestions and discussions. Many thanks to Haopeng Li, Mingyue Li, and Zhanyu Ma, who helped me a lot during my research.
    [Show full text]
  • Respiratory Motion Compensation Using Diaphragm Tracking for Cone-Beam C-Arm CT: a Simulation and a Phantom Study
    Hindawi Publishing Corporation International Journal of Biomedical Imaging Volume 2013, Article ID 520540, 10 pages http://dx.doi.org/10.1155/2013/520540 Research Article Respiratory Motion Compensation Using Diaphragm Tracking for Cone-Beam C-Arm CT: A Simulation and a Phantom Study Marco Bögel,1 Hannes G. Hofmann,1 Joachim Hornegger,1 Rebecca Fahrig,2 Stefan Britzen,3 and Andreas Maier3 1 Pattern Recognition Lab, Friedrich-Alexander-University Erlangen-Nuremberg, 91058 Erlangen, Germany 2 Department of Radiology, Lucas MRS Center, Stanford University, Palo Alto, CA 94304, USA 3 Siemens AG, Healthcare Sector, 91301 Forchheim, Germany Correspondence should be addressed to Andreas Maier; [email protected] Received 21 February 2013; Revised 13 May 2013; Accepted 15 May 2013 Academic Editor: Michael W. Vannier Copyright © 2013 Marco Bogel¨ et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Long acquisition times lead to image artifacts in thoracic C-arm CT. Motion blur caused by respiratory motion leads to decreased image quality in many clinical applications. We introduce an image-based method to estimate and compensate respiratory motion in C-arm CT based on diaphragm motion. In order to estimate respiratory motion, we track the contour of the diaphragm in the projection image sequence. Using a motion corrected triangulation approach on the diaphragm vertex, we are able to estimate a motion signal. The estimated motion signal is used to compensate for respiratory motion in the target region, for example, heart or lungs.
    [Show full text]
  • Video Coding Standards
    Module 8 Video Coding Standards Version 2 ECE IIT, Kharagpur Lesson 23 MPEG-1 standards Version 2 ECE IIT, Kharagpur Lesson objectives At the end of this lesson, the students should be able to : 1. Enlist the major video coding standards 2. State the basic objectives of MPEG-1 standard. 3. Enlist the set of constrained parameters in MPEG-1 4. Define the I- P- and B-pictures 5. Present the hierarchical data structure of MPEG-1 6. Define the macroblock modes supported by MPEG-1 23.0 Introduction In lesson 21 and lesson 22, we studied how to perform motion estimation and thereby temporally predict the video frames to exploit significant temporal redundancies present in the video sequence. The error in temporal prediction is encoded by standard transform domain techniques like the DCT, followed by quantization and entropy coding to exploit the spatial and statistical redundancies and achieve significant video compression. The video codecs therefore follow a hybrid coding structure in which DPCM is adopted in temporal domain and DCT or other transform domain techniques in spatial domain. Efforts to standardize video data exchange via storage media or via communication networks are actively in progress since early 1980s. A number of international video and audio standardization activities started within the International Telephone Consultative Committee (CCITT), followed by the International Radio Consultative Committee (CCIR), and the International Standards Organization / International Electrotechnical Commission (ISO/IEC). An experts group, known as the Motion Pictures Expects Group (MPEG) was established in 1988 in the framework of the Joint ISO/IEC Technical Committee with an objective to develop standards for coded representation of moving pictures, associated audio, and their combination for storage and retrieval of digital media.
    [Show full text]
  • 11.2 Motion Estimation and Motion Compensation 421
    11.2 Motion Estimation and Motion Compensation 421 vertical component to the enhancement filter, making the overall filter separable with 3 3 support. × 11.2 MOTION ESTIMATION AND MOTION COMPENSATION Motion compensation (MC) is very useful in video filtering to remove noise and enhance signal. It is useful since it allows the filter or coder to process through the video on a path of near-maximum correlation based on following motion trajectories across the frames making up the image sequence or video. Motion compensation is also employed in all distribution-quality video coding formats, since it is able to achieve the smallest prediction error, which is then easier to code. Motion can be characterized in terms of either a velocity vector v or displacement vector d and is used to warp a reference frame onto a target frame. Motion estimation is used to obtain these displacements, one for each pixel in the target frame. Several methods of motion estimation are commonly used: • Block matching • Hierarchical block matching • Pel-recursive motion estimation • Direct optical flow methods • Mesh-matching methods Optical flow is the apparent displacement vector field d .d1,d2/ we get from setting (i.e., forcing) equality in the so-called constraint equationD x.n1,n2,n/ x.n1 d1,n2 d2,n 1/. (11.2–1) D − − − All five approaches start from this basic equation, which is really just an ide- alization. Departures from the ideal are caused by the covering and uncovering of objects in the viewed scene, lighting variation both in time and across the objects in the scene, movement toward or away from the camera, as well as rotation about an axis (i.e., 3-D motion).
    [Show full text]