Comparing a Spatial Extension of ICTCP Color Representation with S-CIELAB and Other Recent Color Metrics for HDR and WCG Quality Assessment

Total Page:16

File Type:pdf, Size:1020Kb

Comparing a Spatial Extension of ICTCP Color Representation with S-CIELAB and Other Recent Color Metrics for HDR and WCG Quality Assessment https://doi.org/10.2352/ISSN.2470-1173.2020.15.COLOR-162 © 2020, Society for Imaging Science and Technology Comparing a spatial extension of ICTCP color representation with S-CIELAB and other recent color metrics for HDR and WCG quality assessment Anustup Choudhury, Scott Daly; Dolby Laboratories Inc.; Sunnyvale, CA, USA Abstract of lower spatial frequencies and non-contiguous regions, while Content created in High Dynamic Range (HDR) and Wide the natural imagery has much higher frequencies, masking due Color Gamut (WCG) is becoming more ubiquitous, driving the to textures, as well contiguous color region effects and gradients. need for reliable tools for evaluating the quality across the imag- Furthermore, while there has been significant evaluation for color ing ecosystem. One of the simplest techniques to measure the differences in complex (natural and civilized) imagery [2, 3], most quality of any video system is to measure the color errors. The tra- of this evaluation has been dominated by still images as opposed ditional color difference metrics such as DE00 and the newer HDR to video, and has been almost entirely limited to Standard Dy- specific metrics such as DEZ and DEITP compute color difference namic Range (SDR) content. on a pixel-by-pixel basis which do not account for the spatial ef- For this work, we evaluated several color difference met- fects (optical) and active processing (neural) done by the human rics on four publicly available HDR databases consisting of com- visual system. In this work, we improve upon the per-pixel DEITP plex images with various distortions, along with the correspond- color difference metric by performing a spatial extension simi- ing subjective scores. The different databases focus on differ- lar to what was done during the design of S-CIELAB. We quan- ent distortions and the aggregation of these covers a wide vari- tified the performance using four standard evaluation procedures ety of both luminance and chromatic distortions. Some of these on four publicly available HDR and WCG image databases and distortions result from tone-mapping and gamut mapping opera- found that the proposed metric results in a marked improvement tions, which tend to be dominated by lower frequencies due to with subjective scores over existing per-pixel color difference met- shallow gradients and easily-visible regions, but also can con- rics. tain step edge artifacts (i.e., having a 1/f spectrum). Other dis- tortions include higher frequency distortions resulting from com- Introduction pression artifacts by various compression schemes such as JPEG, Millions of devices now support High Dynamic Range JPEG-XT, JPEG2000, and HEVC. These typically include ring- (HDR) and Wide Color Gamut (WCG) content. Display design, ing around sharp edges (‘mosquito noise’) and visible transform video processing algorithm design, system format development block boundaries (‘blocking’). While perceptually dominated by and comparison across products all require being able to evalu- luminance distortions, these also contain chromatic distortions ate the quality of HDR images in a perceptually relevant manner. due to chroma subsampling and different processes acting on Y, There is a vast body of literature on quality metrics relating to Cr and Cb signals. Image statistics play a key role in imaging image data compression for traditional standard dynamic range product design, as the era of displays being able to expect a certain (SDR) images and a steadily increasing amount of research on class of imagery (e.g., optically-captured, or text, or computer- HDR [1], but these are almost exclusively based exclusively on generated) has given way to systems being used for all types of the luminance component. However, color distortions are also imagery. Some key aspects of image statistics have been studies very important to assess because they are increasingly likely in for SDR, such as the 1= f N spatial frequency power spectra, the HDR1 systems, due to the additional gamut and tone mapping log-normal luminance histograms, and the principal components operations that are required to convert the source color volume to in descending variances of an achromatic and two uncorrelated a typically reduced display color volume. Most commonly used chromatic components. However, the same statistics for HDR color difference metrics have been designed to measure differ- images are less well understood, and even for SDR these statis- ences between simple test patches, as opposed to natural (com- tics do not take into account characteristics such as as texture vs. plex2) imagery. Typically, the size of the test patch does not smooth gradients, mixed illumination, frequency of emissive light match the size of the objects in an image, which are often sub- sources, depth of field, etc [4]. Consequently, to allow for ro- stantially smaller in terms of visual degrees (by a factor of 1/300 bustness, it is desirable to evaluate as many images as possible. in consumer TV applications). Other key differences between test Toward that goal, this work includes a total of 46 source images patches and complex imagery is that test patches typically consist and a total of 532 distorted images as evaluated by 94 observers across all four databases. 1 Since most HDR systems are also WCG, we will use the term ‘HDR’ In addition to the commonly used (DE00) color difference to include both types of advances metric, we compare several recent metrics derived for HDR ap- 2There are specific categories of imagery such as natural (i.e., no human-made objects), civilized (including human-made objects), real- plications: DEZ based on the Jzazbz color space, and DEITP based world (optically captured), synthetic (computer generated) so in this paper on the ICTCP color space. While the CIE L*a*b*-based met- we will use the term ‘complex imagery’ to include all cases rics have been shown to perform well for many SDR applications IS&T International Symposium on Electronic Imaging 2020 Color Imaging: Displaying, Processing, Hardcopy, and Applications 162-1 (product surface colors, graphic arts printing), they are known to ing above threshold (9 luminance steps from white to black and have significant problems with new image characteristics enabled typically less than 9 steps from neutral to the maximum satura- by HDR, such as shadow detail spanning several log units, emis- tion) and consequently under-predicts threshold visibility [15]. sive colors, specular reflections, scenes of mixed illumination, and So, it is a good candidate to describe as a macro-uniform color interscene adaptations (e.g., from scene changes like going into a space. There is current debate on whether the micro-uniform or cave, or turning on a light source). One source of the CIE L*a*b* macro-uniform color space will better predict the kinds of color problems is the cube-root based nonlinearity forming the back- distortions in complex imagery and of practical interest to busi- bone of L*, since it has the inability to handle the continuously ness. One goal of this paper is to see how effective the appearance differing non-linearities such as log behavior (Weber’s Law) for and threshold-based approaches are in predicting the kinds of dis- greater than 200cd=m2, the square-root behavior (Rose-Devries tortions in the databases. To quantify the performance of the dif- law) for less than 0:001cd=m2, and the continuum of changes be- ferent color difference metrics, we use four standard performance tween these extremes. On the other hand, CIE L*a*b* is vetted by evaluation procedures – Root Mean Square Error, Pearson Linear standards and has a long history of use, with many experts familiar Correlation Coefficient, Spearman Rank-Order Correlation Coef- with how to apply it. While we have omitted DE94 in this anal- ficient and Outlier Ratio. ysis, it was found to perform worse than DE00 in a HDR-WCG Typical color difference metrics are pixel-based operations study [5]. and results are shown on test patches. In our previous work [5], Both of the new HDR metrics should have better perfor- we measured the performance of various color difference metrics mance over the larger luminance ranges common to HDR, as on a database of natural HDR/WCG images. However, apply- they have been derived from the behavior of the CSF (contrast ing these metrics on a pixel-by-pixel basis, as done with DEZ and sensitivity function 6= spatial MTF) across a large range of light DEITP, without accounting for spatial information from neighbor- adaptation [6, 7], whereas the L* essentially models a single fixed ing pixels does not really mimic the Human Visual System (HVS). adaptation level. In addition, they are both physiologically more With such approaches, for the same magnitude of color distortion, realistic models since they apply a nonlinearity to the signals for a single pixel can have as large of an effect as a large image region the actual known L, M, S, cones in the retina, as opposed to the if a maximum error criteria is used. Averaging the results across psychophysically-derived X, Y, Z color matching signals as do an entire image can be used to avoid this kind of mis-assessment, the CIE L*a*b*-based metrics (which were developed before the but may hide large errors that span only a few pixels. Consider known L, M, S responses were measured). Working with LMS two cases of distortion. One is a 100x100 contiguous pixel re- cone responses allows for chromatic adaptation to be modeled gion while the other has the same number of pixels (10,000) and without the well-known ‘Wrong Von-Kries Adaptation’ distor- magnitude but the pixels are scattered across the image individu- tions that occur when working on the XYZ signals [8].
Recommended publications
  • Creating 4K/UHD Content Poster
    Creating 4K/UHD Content Colorimetry Image Format / SMPTE Standards Figure A2. Using a Table B1: SMPTE Standards The television color specification is based on standards defined by the CIE (Commission 100% color bar signal Square Division separates the image into quad links for distribution. to show conversion Internationale de L’Éclairage) in 1931. The CIE specified an idealized set of primary XYZ SMPTE Standards of RGB levels from UHDTV 1: 3840x2160 (4x1920x1080) tristimulus values. This set is a group of all-positive values converted from R’G’B’ where 700 mv (100%) to ST 125 SDTV Component Video Signal Coding for 4:4:4 and 4:2:2 for 13.5 MHz and 18 MHz Systems 0mv (0%) for each ST 240 Television – 1125-Line High-Definition Production Systems – Signal Parameters Y is proportional to the luminance of the additive mix. This specification is used as the color component with a color bar split ST 259 Television – SDTV Digital Signal/Data – Serial Digital Interface basis for color within 4K/UHDTV1 that supports both ITU-R BT.709 and BT2020. 2020 field BT.2020 and ST 272 Television – Formatting AES/EBU Audio and Auxiliary Data into Digital Video Ancillary Data Space BT.709 test signal. ST 274 Television – 1920 x 1080 Image Sample Structure, Digital Representation and Digital Timing Reference Sequences for The WFM8300 was Table A1: Illuminant (Ill.) Value Multiple Picture Rates 709 configured for Source X / Y BT.709 colorimetry ST 296 1280 x 720 Progressive Image 4:2:2 and 4:4:4 Sample Structure – Analog & Digital Representation & Analog Interface as shown in the video ST 299-0/1/2 24-Bit Digital Audio Format for SMPTE Bit-Serial Interfaces at 1.5 Gb/s and 3 Gb/s – Document Suite Illuminant A: Tungsten Filament Lamp, 2854°K x = 0.4476 y = 0.4075 session display.
    [Show full text]
  • Khronos Data Format Specification
    Khronos Data Format Specification Andrew Garrard Version 1.2, Revision 1 2019-03-31 1 / 207 Khronos Data Format Specification License Information Copyright (C) 2014-2019 The Khronos Group Inc. All Rights Reserved. This specification is protected by copyright laws and contains material proprietary to the Khronos Group, Inc. It or any components may not be reproduced, republished, distributed, transmitted, displayed, broadcast, or otherwise exploited in any manner without the express prior written permission of Khronos Group. You may use this specification for implementing the functionality therein, without altering or removing any trademark, copyright or other notice from the specification, but the receipt or possession of this specification does not convey any rights to reproduce, disclose, or distribute its contents, or to manufacture, use, or sell anything that it may describe, in whole or in part. This version of the Data Format Specification is published and copyrighted by Khronos, but is not a Khronos ratified specification. Accordingly, it does not fall within the scope of the Khronos IP policy, except to the extent that sections of it are normatively referenced in ratified Khronos specifications. Such references incorporate the referenced sections into the ratified specifications, and bring those sections into the scope of the policy for those specifications. Khronos Group grants express permission to any current Promoter, Contributor or Adopter member of Khronos to copy and redistribute UNMODIFIED versions of this specification in any fashion, provided that NO CHARGE is made for the specification and the latest available update of the specification for any version of the API is used whenever possible.
    [Show full text]
  • Yasser Syed & Chris Seeger Comcast/NBCU
    Usage of Video Signaling Code Points for Automating UHD and HD Production-to-Distribution Workflows Yasser Syed & Chris Seeger Comcast/NBCU Comcast TPX 1 VIPER Architecture Simpler Times - Delivering to TVs 720 1920 601 HD 486 1080 1080i 709 • SD - HD Conversions • Resolution, Standard Dynamic Range and 601/709 Color Spaces • 4:3 - 16:9 Conversions • 4:2:0 - 8-bit YUV video Comcast TPX 2 VIPER Architecture What is UHD / 4K, HFR, HDR, WCG? HIGH WIDE HIGHER HIGHER DYNAMIC RESOLUTION COLOR FRAME RATE RANGE 4K 60p GAMUT Brighter and More Colorful Darker Pixels Pixels MORE FASTER BETTER PIXELS PIXELS PIXELS ENABLED BY DOLBYVISION Comcast TPX 3 VIPER Architecture Volume of Scripted Workflows is Growing Not considering: • Live Events (news/sports) • Localized events but with wider distributions • User-generated content Comcast TPX 4 VIPER Architecture More Formats to Distribute to More Devices Standard Definition Broadcast/Cable IPTV WiFi DVDs/Files • More display devices: TVs, Tablets, Mobile Phones, Laptops • More display formats: SD, HD, HDR, 4K, 8K, 10-bit, 8-bit, 4:2:2, 4:2:0 • More distribution paths: Broadcast/Cable, IPTV, WiFi, Laptops • Integration/Compositing at receiving device Comcast TPX 5 VIPER Architecture Signal Normalization AUTOMATED LOGIC FOR CONVERSION IN • Compositing, grading, editing SDR HLG PQ depends on proper signal BT.709 BT.2100 BT.2100 normalization of all source files (i.e. - Still Graphics & Titling, Bugs, Tickers, Requires Conversion Native Lower-Thirds, weather graphics, etc.) • ALL content must be moved into a single color volume space. Normalized Compositing • Transformation from different Timeline/Switcher/Transcoder - PQ-BT.2100 colourspaces (BT.601, BT.709, BT.2020) and transfer functions (Gamma 2.4, PQ, HLG) Convert Native • Correct signaling allows automation of conversion settings.
    [Show full text]
  • How Close Is Close Enough? Specifying Colour Tolerances for Hdr and Wcg Displays
    HOW CLOSE IS CLOSE ENOUGH? SPECIFYING COLOUR TOLERANCES FOR HDR AND WCG DISPLAYS Jaclyn A. Pytlarz, Elizabeth G. Pieri Dolby Laboratories Inc., USA ABSTRACT With a new high-dynamic-range (HDR) and wide-colour-gamut (WCG) standard defined in ITU-R BT.2100 (1), display and projector manufacturers are racing to extend their visible colour gamut by brightening and widening colour primaries. The question is: how close is close enough? Having this answer is increasingly important for both consumer and professional display manufacturers who strive to balance design trade-offs. In this paper, we present “ground truth” visible colour differences from a psychophysical experiment using HDR laser cinema projectors with near BT.2100 colour primaries up to 1000 cd/m2. We present our findings, compare colour difference metrics, and propose specifying colour tolerances for HDR/WCG displays using the ΔICTCP (2) metric. INTRODUCTION AND BACKGROUND From initial display design to consumer applications, measuring colour differences is a vital component of the imaging pipeline. Now that the industry has moved towards displays with higher dynamic range as well as wider, more saturated colours, no standardized method of measuring colour differences exists. In display calibration, aside from metamerism effects, it is crucial that the specified tolerances align with human perception. Otherwise, one of two undesirable situations might result: first, tolerances are too large and calibrated displays will not appear to visually match; second, tolerances are unnecessarily tight and the calibration process becomes uneconomic. The goal of this paper is to find a colour difference measurement metric for HDR/WCG displays that balances the two and closely aligns with human vision.
    [Show full text]
  • Khronos Data Format Specification
    Khronos Data Format Specification Andrew Garrard Version 1.3.1 2020-04-03 1 / 281 Khronos Data Format Specification License Information Copyright (C) 2014-2019 The Khronos Group Inc. All Rights Reserved. This specification is protected by copyright laws and contains material proprietary to the Khronos Group, Inc. It or any components may not be reproduced, republished, distributed, transmitted, displayed, broadcast, or otherwise exploited in any manner without the express prior written permission of Khronos Group. You may use this specification for implementing the functionality therein, without altering or removing any trademark, copyright or other notice from the specification, but the receipt or possession of this specification does not convey any rights to reproduce, disclose, or distribute its contents, or to manufacture, use, or sell anything that it may describe, in whole or in part. This version of the Data Format Specification is published and copyrighted by Khronos, but is not a Khronos ratified specification. Accordingly, it does not fall within the scope of the Khronos IP policy, except to the extent that sections of it are normatively referenced in ratified Khronos specifications. Such references incorporate the referenced sections into the ratified specifications, and bring those sections into the scope of the policy for those specifications. Khronos Group grants express permission to any current Promoter, Contributor or Adopter member of Khronos to copy and redistribute UNMODIFIED versions of this specification in any fashion, provided that NO CHARGE is made for the specification and the latest available update of the specification for any version of the API is used whenever possible.
    [Show full text]
  • KONA Series Capture, Display, Convert
    KONA Series Capture, Display, Convert Installation and Operation Manual Version 16.1 Published July 14, 2021 Notices Trademarks AJA® and Because it matters.® are registered trademarks of AJA Video Systems, Inc. for use with most AJA products. AJA™ is a trademark of AJA Video Systems, Inc. for use with recorder, router, software and camera products. Because it matters.™ is a trademark of AJA Video Systems, Inc. for use with camera products. Corvid Ultra®, lo®, Ki Pro®, KONA®, KUMO®, ROI® and T-Tap® are registered trademarks of AJA Video Systems, Inc. AJA Control Room™, KiStor™, Science of the Beautiful™, TruScale™, V2Analog™ and V2Digital™ are trademarks of AJA Video Systems, Inc. All other trademarks are the property of their respective owners. Copyright Copyright © 2021 AJA Video Systems, Inc. All rights reserved. All information in this manual is subject to change without notice. No part of the document may be reproduced or transmitted in any form, or by any means, electronic or mechanical, including photocopying or recording, without the express written permission of AJA Video Systems, Inc. Contacting AJA Support When calling for support, have all information at hand prior to calling. To contact AJA for sales or support, use any of the following methods: Telephone +1.530.271.3190 FAX +1.530.271.3140 Web https://www.aja.com Support Email [email protected] Sales Email [email protected] KONA Capture, Display, Convert v16.1 2 www.aja.com Contents Notices . .2 Trademarks . 2 Copyright . 2 Contacting AJA Support . 2 Chapter 1 – Introduction . .5 Overview. .5 KONA Models Covered in this Manual .
    [Show full text]
  • User Requirements for Video Monitors in Television Production
    TECH 3320 USER REQUIREMENTS FOR VIDEO MONITORS IN TELEVISION PRODUCTION VERSION 4.1 Geneva September 2019 This page and several other pages in the document are intentionally left blank. This document is paginated for two sided printing Tech 3320 v4.1 User requirements for Video Monitors in Television Production Conformance Notation This document contains both normative text and informative text. All text is normative except for that in the Introduction, any § explicitly labeled as ‘Informative’ or individual paragraphs which start with ‘Note:’. Normative text describes indispensable or mandatory elements. It contains the conformance keywords ‘shall’, ‘should’ or ‘may’, defined as follows: ‘Shall’ and ‘shall not’: Indicate requirements to be followed strictly and from which no deviation is permitted in order to conform to the document. ‘Should’ and ‘should not’: Indicate that, among several possibilities, one is recommended as particularly suitable, without mentioning or excluding others. OR indicate that a certain course of action is preferred but not necessarily required. OR indicate that (in the negative form) a certain possibility or course of action is deprecated but not prohibited. ‘May’ and ‘need not’: Indicate a course of action permissible within the limits of the document. Informative text is potentially helpful to the user, but it is not indispensable, and it does not affect the normative text. Informative text does not contain any conformance keywords. Unless otherwise stated, a conformant implementation is one which includes all mandatory provisions (‘shall’) and, if implemented, all recommended provisions (‘should’) as described. A conformant implementation need not implement optional provisions (‘may’) and need not implement them as described.
    [Show full text]
  • Application of Contrast Sensitivity Functions in Standard and High Dynamic Range Color Spaces
    https://doi.org/10.2352/ISSN.2470-1173.2021.11.HVEI-153 This work is licensed under the Creative Commons Attribution 4.0 International License. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/. Color Threshold Functions: Application of Contrast Sensitivity Functions in Standard and High Dynamic Range Color Spaces Minjung Kim, Maryam Azimi, and Rafał K. Mantiuk Department of Computer Science and Technology, University of Cambridge Abstract vision science. However, it is not obvious as to how contrast Contrast sensitivity functions (CSFs) describe the smallest thresholds in DKL would translate to thresholds in other color visible contrast across a range of stimulus and viewing param- spaces across their color components due to non-linearities such eters. CSFs are useful for imaging and video applications, as as PQ encoding. contrast thresholds describe the maximum of color reproduction In this work, we adapt our spatio-chromatic CSF1 [12] to error that is invisible to the human observer. However, existing predict color threshold functions (CTFs). CTFs describe detec- CSFs are limited. First, they are typically only defined for achro- tion thresholds in color spaces that are more commonly used in matic contrast. Second, even when they are defined for chromatic imaging, video, and color science applications, such as sRGB contrast, the thresholds are described along the cardinal dimen- and YCbCr. The spatio-chromatic CSF model from [12] can sions of linear opponent color spaces, and therefore are difficult predict detection thresholds for any point in the color space and to relate to the dimensions of more commonly used color spaces, for any chromatic and achromatic modulation.
    [Show full text]
  • Installation and Operation Manual
    Io IP Transport, Capture, Display Installation and Operation Manual Version 16.1 Published July 14, 2021 Notices Trademarks AJA® and Because it matters.® are registered trademarks of AJA Video Systems, Inc. for use with most AJA products. AJA™ is a trademark of AJA Video Systems, Inc. for use with recorder, router, software and camera products. Because it matters.™ is a trademark of AJA Video Systems, Inc. for use with camera products. Corvid Ultra®, lo®, Ki Pro®, KONA®, KUMO®, ROI® and T-Tap® are registered trademarks of AJA Video Systems, Inc. AJA Control Room™, KiStor™, Science of the Beautiful™, TruScale™, V2Analog™ and V2Digital™ are trademarks of AJA Video Systems, Inc. All other trademarks are the property of their respective owners. Copyright Copyright © 2021 AJA Video Systems, Inc. All rights reserved. All information in this manual is subject to change without notice. No part of the document may be reproduced or transmitted in any form, or by any means, electronic or mechanical, including photocopying or recording, without the express written permission of AJA Video Systems, Inc. Contacting AJA Support When calling for support, have all information at hand prior to calling. To contact AJA for sales or support, use any of the following methods: Telephone +1.530.271.3190 FAX +1.530.271.3140 Web https://www.aja.com Support Email [email protected] Sales Email [email protected] Io IP Transport, Capture, Display v16.1 2 www.aja.com Contents Notices . .2 Trademarks . 2 Copyright . 2 Contacting AJA Support . 2 Chapter 1 – Introduction . .5 Overview. .5 Io IP Features . 5 AJA Software & Utilities .
    [Show full text]
  • Common Metadata Ref: TR-META-CM Version: 2.7 DRAFT DRAFT Date: September 23,2018
    Common Metadata Ref: TR-META-CM Version: 2.7 DRAFT DRAFT Date: September 23,2018 Common Metadata ‘md’ namespace Motion Picture Laboratories, Inc. i Common Metadata Ref: TR-META-CM Version: 2.7 DRAFT DRAFT Date: September 23,2018 CONTENTS 1 Introduction .............................................................................................................. 1 1.1 Overview of Common Metadata ....................................................................... 1 1.2 Document Organization .................................................................................... 1 1.3 Document Notation and Conventions ............................................................... 2 1.3.1 XML Conventions ...................................................................................... 2 1.3.2 General Notes ........................................................................................... 3 1.4 Normative References ...................................................................................... 4 1.5 Informative References..................................................................................... 7 1.6 Best Practices for Maximum Compatibility ........................................................ 8 2 Identifiers ............................................................................................................... 10 2.1 Identifier Structure .......................................................................................... 10 2.1.1 ID Simple Types .....................................................................................
    [Show full text]
  • SMPTE ST 2094 and Dynamic Metadata
    SMPTE Standards Update SMPTE Professional Development Academy – Enabling Global Education SMPTE Standards Webcast Series SMPTE Professional Development Academy – Enabling Global Education SMPTE ST 2094 and Dynamic Metadata Lars Borg Principal Scientist Adobe linkedin: larsborg My first TV © 2017 by the Society of Motion Picture & Television Engineers®, Inc. (SMPTE®) SMPTE Standards Update Webcasts Professional Development Academy Enabling Global Education • Series of quarterly 1-hour online (this is is 90 minutes), interactive webcasts covering select SMPTE standards • Free to everyone • Sessions are recorded for on-demand viewing convenience SMPTE.ORG and YouTube © 2017 by the Society of Motion Picture & Television Engineers®, Inc. (SMPTE®) © 2016• Powered by SMPTE® Professional Development Academy • Enabling Global Education • www.smpte.org © 2017 • Powered by SMPTE® Professional Development Academy | Enabling Global Education • www.smpte.org SMPTE Standards Update SMPTE Professional Development Academy – Enabling Global Education Your Host Professional Development Academy • Joel E. Welch Enabling Global Education • Director of Education • SMPTE © 2017 by the Society of Motion Picture & Television Engineers®, Inc. (SMPTE®) 3 Views and opinions expressed during this SMPTE Webcast are those of the presenter(s) and do not necessarily reflect those of SMPTE or SMPTE Members. This webcast is presented for informational purposes only. Any reference to specific companies, products or services does not represent promotion, recommendation, or endorsement by SMPTE © 2017 • Powered by SMPTE® Professional Development Academy | Enabling Global Education • www.smpte.org SMPTE Standards Update SMPTE Professional Development Academy – Enabling Global Education Today’s Guest Speaker Professional Development Academy Lars Borg Enabling Global Education Principal Scientist in Digital Video and Audio Engineering Adobe © 2017 by the Society of Motion Picture & Television Engineers®, Inc.
    [Show full text]
  • DELTA-12G1c-Hmi10-Elp Dual-Channel 4K60 Mixed HMDI™ Capture and SDI I/O Card
    DELTA-12G1c-hmi10-elp Dual-channel 4K60 mixed HMDI™ capture and SDI I/O card The DELTA-12G1c-hmi10-elp mixed interface model is a dual-channel 4K60 card, capable of simultaneous capturing UHDTV (4K60) on HDMI™ and SDI. The SDI interfaces are bidirectional, meaning that each port can be configured by the software to act as an input or as an output. The card hosts four 3G/HD/SD-SDI ports, with one of them also supporting 6G and 12G-SDI. The HDMI 2.0b (600MHz TMDS) receiver along with the 12G or 4x 3G-SDI I/Os offer the best of both standards for 4K60 capture in both HDMI and SDI; and 4K60 playout in SDI, on a single compact, low-profile PCIe card. Designed for high-performance applications and HDR workflows, the DELTA-12G1c-hmi10-elp includes on-board color processing features (color space converters, chroma resampler and planar buffer packing), proxy scalers, embedded audio and HDR metadata support. Hardware UHDTV SDI interfaces Low-profile, half-length PCIe form factor Support of 1x UHDTV capture or playout (SMPTE formats) : PCIe Gen-3 x8 electrical interface Up to 4K60 in Single-Link 12G - ST 2082-10 1x HDMI 2.0b (600MHz TMDS clock) input Up to 4K60 in Quad-Link 3G HDMI Type A connector o 2 Sample Interleave (2SI) - ST 425-5 4x SDI bi-directional I/Os ports o Square Division (SQD) quad split - 4x 3G A or B- o Port 1 : 12G/6G/3G/HD/SD DL. o Port 2-4 : 3G/HD/SD Up to 4K30 in Single-Link 6G - ST 2081-10 HD-BNC 75Ohm high-frequency connectors Up to 4K30 in Dual-Link 3G – ST 425-3 2GB DDR-4 on-board buffer Reference input for
    [Show full text]