A Color Temperature-Based High-Speed Decolorization: an Empirical Approach for Tone Mapping Applications

Total Page:16

File Type:pdf, Size:1020Kb

A Color Temperature-Based High-Speed Decolorization: an Empirical Approach for Tone Mapping Applications 1 A color temperature-based high-speed decolorization: an empirical approach for tone mapping applications Prasoon Ambalathankandy ID˙ , Yafei Ou ID˙ , and Masayuki Ikebe ID˙ Abstract—Grayscale images are fundamental to many image into global and local methods. Global methods can define processing applications like data compression, feature extraction, only one conversion function for all pixels, and most of printing and tone mapping. However, some image information these methods use all pixels in the image to determine the is lost when converting from color to grayscale. In this paper, we propose a light-weight and high-speed image decolorization function. On the other hand, local ones process the target method based on human perception of color temperatures. from neighboring pixels in the same way as a spatial filter, Chromatic aberration results from differential refraction of light the function is different for each pixel. However, both types depending on its wavelength. It causes some rays corresponding of methods face the issue of calculation cost, which comes to cooler colors (like blue, green) to converge before the warmer from optimization iterations or spatial filter processing. colors (like red, orange). This phenomena creates a perception of warm colors “advancing” toward the eye, while the cool colors We have developed a fast decolorization method that reflects to be “receding” away. In this proposed color to gray conversion the perception of warm and cool colors which is well known model, we implement a weighted blending function to combine in psychophysics studies [1]. Colors are arranged according to red (perceived warm) and blue (perceived cool) channel. Our their wavelengths on the color wheel, the ones with longest main contribution is threefold: First, we implement a high-speed wavelengths are on the right side of the wheel and are known color processing method using exact pixel by pixel processing, and we report a 5:7× speed up when compared to other as warm colors, as they evoke warmth. These hues include new algorithms. Second, our optimal color conversion method shades of red, yellow, and orange. On the other hand, green, produces luminance in images that are comparable to other blue and violet which have shorter wavelengths are placed state of the art methods which we quantified using the objective on the left side of the color wheel, and are perceived as cool metrics (E-score and C2G-SSIM) and a subjective user study. colors. The color of an object in a scene affects our perception Third, we demonstrate that an effective luminance distribution can be achieved using our algorithm by using global and local of its apparent depth and this phenomenon has been exploited tone mapping applications. by many artists. This optical illusion has been studied by psychologists, and early researchers explored color-depth re- Index Terms—Color temperature, chromatic aberration, de- colorization, luminance, pre-processing, RGB, tonemap. lationship. One of the widely accepted theory explains this phenomenon is due to fact that shorter wavelengths of visible light are refracted more than longer wavelengths [2]. In other I. INTRODUCTION words, an equidistant source of different wavelengths cannot Grayscale channels, which reflect image luminance, are be focused simultaneously onto our retina. This phenomena used for various applications such as printing, tone mapping, is called as chromatic aberration and we discuss it in detail data compression, and feature extraction. Thus, obtaining lu- in section III-A. In our decolorization method, we combine minance along with human perception has a key role for decol- the idea of warm/cool colors with the Helmholtz-Kohlrausch orization, which converts RGB channels to high-quality gray effect [3]. On that account, we make two assumptions, which arXiv:2108.13656v1 [eess.IV] 31 Aug 2021 ones. For example, High Dynamic Range (HDR) compression are: (i) warm colors (mainly including R) are lighter than Y is ideally performed by tone mapping the luminance channel of YCbCr and (ii) mixed colors are darker than the Y or L of for the lower computational and memory cost. However, CIE with the same luminance. To satisfy these assumptions, applying well-known luminance channels such as Y of YCbCr we use a weighted blending of RGB channels and remap them or V of HSV does not guarantee appropriate tone mapping, to warm/cool colors on the luminance channel. Following are as these channels do not reflect human perceptions. Therefore, our main contributions: decolorization has gathered considerable attention and various • We achieve high-speed color mapping by exact pixel-by- sophisticated methods to achieve perceptual decolorization pixel processing. have recently been proposed. These methods can be classified • We obtain luminance comparable to that of optimization- Prasoon Ambalathankandy and Yafei Ou are with the Research Center based methods. For Integrated Quantum Electronics, Hokkaido University, Sapporo 060- 0813, Japan, and also with the Graduate School of Information Science and • We demonstrate effective luminance distribution by per- Technology, Hokkaido University, Sapporo 060-0814, Japan. forming objective evaluations and a subjective user study. Masayuki Ikebe is with the Research Center For Integrated Quan- tum Electronics, Hokkaido University, Sapporo 060-0813, Japan (e-mail: There are many well defined methods to convert any color [email protected]). image to a grayscale image. An effortless procedure is to 2 cool warm Luminance advance HVS based recede Original Image color mapping This work YCbCr: Lum. Fig. 1: Main concept of our decolorization method: Human perception of warm and cool colors. Warm colors “advance” toward the eye, while cool colors “recede”. In this work we are able to accurately reflect the human perception of color temperatures, whereas this phenomenon is non-existent in conventional YCbCr color space. In the global methods category, Gooch et al. [4] proposed a global decolorization algorithm that can be implemented by solving the optimization problem for all image pixels. Then, Kim et al. [7] aimed at high-speed processing by simplifying Gooch’s method. Smith et al. [6] used unsharp masking and the Original Image Y of YCbCr CIE Lab This work Helmholtz-Kohlrausch effect (H-K effect) model of Nayatani et al. [3]. Nayatani’s model [3] is merely an experimental Fig. 2: Comparison of luminance components obtained using model for the effect of the CIELUV chrominance component YCbCr, CIELAB and our proposed method. for human perception. On the other hand, the method of Lu et al. [8] is focused on converting a color image into a gray image with high contrast. The main advantage of these assign different weights to color channels, in order to have the methods is transformation consistency, i.e., the same color is same luminance in the grayscale image as the original color converted to the same grayscale. However, speed remains a image. For example, in the MATLAB function rgb2gray, problem for these methods. Most of the local methods are it converts any RGB values to grayscale (Gray) values by aimed at speeding up the method of Lu et al [8]. To enhance forming a weighted sum of the R, G, and B components as image contrast, Ancuti et al. [10] adopted the strategy of Gray = 0:2989×R+0:5870×G+0:1140×B. This function using a Laplacian filter and Song et al. [12] used a Sobel operates under an assumption that human visual system is filter. Although local methods are effective in terms of contrast more sensitive to green color. When operating with CIELab emphasis, they are disadvantageous in terms of tone mapping and YUV color spaces, one could directly obtain luminance because conversion consistency is not maintained and it differs channel as the grayscale version of the color image as they from human perception. consider the luminance and color channel to be independent. Recently, some machine learning-based techniques have But, such crude approaches will fail to preserve image contrast also been proposed for image decolorization [13]–[17]. Cai as shown in these examples (see Fig. 1 and Fig. 2). et al. [14] proposed a method, that used the perceptual loss In several real-world image/video processing applications function to pretrain VGG-Net [18]. However, it is difficult like detail enhancement, image matching, and segmentation to control and many of their output images are far from under different illumintion a 1-D grayscale image has to be human perception additionally, the computational cost are obtained from its corresponding 3-D color image. However, high. Processing an image of 256 × 256 size it requires mapping the 3-D color information onto a 1-D grayscale roughly 30 seconds on a single Nvidia GeForceGTX 1080 image while retaining the original contrast and fine details is a GPU. Zhang et al., proposed a CNN framework that combines challenging problem. Additionally, implementing decoloriza- local and global image features [17]. However, their network tion algorithms with a reasonable computational efficiency is framework do not account for exposure features [13]. Lin pivotal for realising their real-time applications. Many studies et al.’s method [16] by utilizing a database of 50 images have been carried out to develop novel decolorization methods. from the Corel dataset produced 50 grayscale images using These mapping methods can be categorized into global [4]– the Color2Gray algorithm [4]. With these 50 input/output [8] and local methods [9]–[12]. In local mapping methods, the image pairs as training examples for their partial differential same color pixel within an image could be mapped into differ- equations-based (PDE)learning system, the learn color2gray ent grayscale values depending on its spatial location. Ideally mapping. The PDE system produced images of comparable this undesirable as such output images may be perceived as quality to [4]. However, for an input image of size n × n their unnatural.
Recommended publications
  • Investigating the Effect of Color Gamut Mapping Quantitatively and Visually
    Rochester Institute of Technology RIT Scholar Works Theses 5-2015 Investigating the Effect of Color Gamut Mapping Quantitatively and Visually Anupam Dhopade Follow this and additional works at: https://scholarworks.rit.edu/theses Recommended Citation Dhopade, Anupam, "Investigating the Effect of Color Gamut Mapping Quantitatively and Visually" (2015). Thesis. Rochester Institute of Technology. Accessed from This Thesis is brought to you for free and open access by RIT Scholar Works. It has been accepted for inclusion in Theses by an authorized administrator of RIT Scholar Works. For more information, please contact [email protected]. Investigating the Effect of Color Gamut Mapping Quantitatively and Visually by Anupam Dhopade A thesis submitted in partial fulfillment of the requirements for the degree of Master of Science in Print Media in the School of Media Sciences in the College of Imaging Arts and Sciences of the Rochester Institute of Technology May 2015 Primary Thesis Advisor: Professor Robert Chung Secondary Thesis Advisor: Professor Christine Heusner School of Media Sciences Rochester Institute of Technology Rochester, New York Certificate of Approval Investigating the Effect of Color Gamut Mapping Quantitatively and Visually This is to certify that the Master’s Thesis of Anupam Dhopade has been approved by the Thesis Committee as satisfactory for the thesis requirement for the Master of Science degree at the convocation of May 2015 Thesis Committee: __________________________________________ Primary Thesis Advisor, Professor Robert Chung __________________________________________ Secondary Thesis Advisor, Professor Christine Heusner __________________________________________ Graduate Director, Professor Christine Heusner __________________________________________ Administrative Chair, School of Media Sciences, Professor Twyla Cummings ACKNOWLEDGEMENT I take this opportunity to express my sincere gratitude and thank all those who have supported me throughout the MS course here at RIT.
    [Show full text]
  • Fast and Stable Color Balancing for Images and Augmented Reality
    Fast and Stable Color Balancing for Images and Augmented Reality Thomas Oskam 1,2 Alexander Hornung 1 Robert W. Sumner 1 Markus Gross 1,2 1 Disney Research Zurich 2 ETH Zurich Abstract This paper addresses the problem of globally balanc- ing colors between images. The input to our algorithm is a sparse set of desired color correspondences between a source and a target image. The global color space trans- formation problem is then solved by computing a smooth Source Image Target Image Color Balanced vector field in CIE Lab color space that maps the gamut of the source to that of the target. We employ normalized ra- dial basis functions for which we compute optimized shape parameters based on the input images, allowing for more faithful and flexible color matching compared to existing RBF-, regression- or histogram-based techniques. Further- more, we show how the basic per-image matching can be Rendered Objects efficiently and robustly extended to the temporal domain us- Tracked Colors balancing Augmented Image ing RANSAC-based correspondence classification. Besides Figure 1. Two applications of our color balancing algorithm. Top: interactive color balancing for images, these properties ren- an underexposed image is balanced using only three user selected der our method extremely useful for automatic, consistent correspondences to a target image. Bottom: our extension for embedding of synthetic graphics in video, as required by temporally stable color balancing enables seamless compositing applications such as augmented reality. in augmented reality applications by using known colors in the scene as constraints. 1. Introduction even for different scenes. With today’s tools this process re- quires considerable, cost-intensive manual efforts.
    [Show full text]
  • Hiding Color Watermarks in Halftone Images Using Maximum-Similarity
    Signal Processing: Image Communication 48 (2016) 1–11 Contents lists available at ScienceDirect Signal Processing: Image Communication journal homepage: www.elsevier.com/locate/image Hiding color watermarks in halftone images using maximum- similarity binary patterns Pedro Garcia Freitas a,n, Mylène C.Q. Farias b, Aletéia P.F. Araújo a a Department of Computer Science, University of Brasília (UnB), Brasília, Brazil b Department of Electrical Engineering, University of Brasília (UnB), Brasília, Brazil article info abstract Article history: This paper presents a halftoning-based watermarking method that enables the embedding of a color Received 3 June 2016 image into binary black-and-white images. To maintain the quality of halftone images, the method maps Received in revised form watermarks to halftone channels using homogeneous dot patterns. These patterns use a different binary 25 August 2016 texture arrangement to embed the watermark. To prevent a degradation of the host image, a max- Accepted 25 August 2016 imization problem is solved to reduce the associated noise. The objective function of this maximization Available online 26 August 2016 problem is the binary similarity measure between the original binary halftone and a set of randomly Keywords: generated patterns. This optimization problem needs to be solved for each dot pattern, resulting in Color embedding processing overhead and a long running time. To overcome this restriction, parallel computing techni- Halftone ques are used to decrease the processing time. More specifically, the method is tested using a CUDA- Color restoration based parallel implementation, running on GPUs. The proposed technique produces results with high Watermarking Enhancement visual quality and acceptable processing time.
    [Show full text]
  • 14. Color Mapping
    14. Color Mapping Jacobs University Visualization and Computer Graphics Lab Recall: RGB color model Jacobs University Visualization and Computer Graphics Lab Data Analytics 691 CMY color model • The CMY color model is related to the RGB color model. •Itsbasecolorsare –cyan(C) –magenta(M) –yellow(Y) • They are arranged in a 3D Cartesian coordinate system. • The scheme is subtractive. Jacobs University Visualization and Computer Graphics Lab Data Analytics 692 Subtractive color scheme • CMY color model is subtractive, i.e., adding colors makes the resulting color darker. • Application: color printers. • As it only works perfectly in theory, typically a black cartridge is added in practice (CMYK color model). Jacobs University Visualization and Computer Graphics Lab Data Analytics 693 CMY color cube • All colors c that can be generated are represented by the unit cube in the 3D Cartesian coordinate system. magenta blue red black grey white cyan yellow green Jacobs University Visualization and Computer Graphics Lab Data Analytics 694 CMY color cube Jacobs University Visualization and Computer Graphics Lab Data Analytics 695 CMY color model Jacobs University Visualization and Computer Graphics Lab Data Analytics 696 CMYK color model Jacobs University Visualization and Computer Graphics Lab Data Analytics 697 Conversion • RGB -> CMY: • CMY -> RGB: Jacobs University Visualization and Computer Graphics Lab Data Analytics 698 Conversion • CMY -> CMYK: • CMYK -> CMY: Jacobs University Visualization and Computer Graphics Lab Data Analytics 699 HSV color model • While RGB and CMY color models have their application in hardware implementations, the HSV color model is based on properties of human perception. • Its application is for human interfaces. Jacobs University Visualization and Computer Graphics Lab Data Analytics 700 HSV color model The HSV color model also consists of 3 channels: • H: When perceiving a color, we perceive the dominant wavelength.
    [Show full text]
  • GS9 Color Management
    Ghostscript 9.21 Color Management Michael J. Vrhel, Ph.D. Artifex Software 7 Mt. Lassen Drive, A-134 San Rafael, CA 94903, USA www.artifex.com Abstract This document provides information about the color architecture in Ghostscript 9.21. The document is suitable for users who wish to obtain accurate color with their output device as well as for developers who wish to customize Ghostscript to achieve a higher level of control and/or interface with a different color management module. Revision 1.6 Artifex Software Inc. www.artifex.com 1 1 Introduction With release 9.0, the color architecture of Ghostscript was updated to primarily use the ICC[1] format for its color management needs. Prior to this release, Ghostscript's color architecture was based heavily upon PostScript[2] Color Management (PCM). This is due to the fact that Ghostscript was designed prior to the ICC format and likely even before there was much thought about digital color management. At that point in time, color management was very much an art with someone adjusting controls to achieve the proper output color. Today, almost all print color management is performed using ICC profiles as opposed to PCM. This fact along with the desire to create a faster, more flexible design was the motivation for the color architectural changes in release 9.0. Since 9.0, several new features and capabilities have been added. As of the 9.21 release, features of the color architecture include: • Easy to interface different CMMs (Color Management Modules) with Ghostscript. • ALL color spaces are defined in terms of ICC profiles.
    [Show full text]
  • Colormap : a Data-Driven Approach and Tool for Mapping Multivariate
    IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, VOL. 24, NO. X, XXXXX 2018 1 ND 1 ColorMap : A Data-Driven Approach and Tool 2 for Mapping Multivariate Data to Color 3 Shenghui Cheng, Wei Xu, Member, IEEE, and Klaus Mueller, Senior Member, IEEE 4 Abstract—A wide variety of color schemes have been devised for mapping scalar data to color. We address the challenge of 5 color-mapping multivariate data. While a number of methods can map low-dimensional data to color, for example, using bilinear 6 or barycentric interpolation for two or three variables, these methods do not scale to higher data dimensions. Likewise, schemes that 7 take a more artistic approach through color mixing and the like also face limits when it comes to the number of variables they can 8 encode. Our approach does not have these limitations. It is data driven in that it determines a proper and consistent color map from first 9 embedding the data samples into a circular interactive multivariate color mapping display (ICD) and then fusing this display with a 10 convex (CIE HCL) color space. The variables (data attributes) are arranged in terms of their similarity and mapped to the ICD’s 11 boundary to control the embedding. Using this layout, the color of a multivariate data sample is then obtained via modified generalized 12 barycentric coordinate interpolation of the map. The system we devised has facilities for contrast and feature enhancement, supports 13 both regular and irregular grids, can deal with multi-field as well as multispectral data, and can produce heat maps, choropleth maps, 14 and diagrams such as scatterplots.
    [Show full text]
  • TONE MAPPING OPTIMIZATION for REAL TIME APPLICATIONS By
    TONE MAPPING OPTIMIZATION FOR REAL TIME APPLICATIONS by Siwei Zhao Submitted in partial fulfilment of the requirements for the degree of Master of Computer Science at Dalhousie University Halifax, Nova Scotia August 2019 © Copyright by Siwei Zhao, 2019 Table of Contents List of Tables .............................................................................................................................v List of Figures .......................................................................................................................... vi Abstract ................................................................................................................................. viii List of Abbreviations Used ....................................................................................................... ix Chapter 1 Introduction ..............................................................................................................1 1.1 Research Problems ......................................................................................................3 1.1.1 Objective Quality Evaluation for HDR Gaming Content ........................................3 1.1.2 Lookup Tables (LUTs) Interpolation In Video Games ............................................4 1.2 Objectives ...................................................................................................................5 1.3 Contributions ..............................................................................................................5 1.4
    [Show full text]
  • Approximate Cross Channel Color Mapping from Sparse Color Correspondences
    2013 IEEE International Conference on Computer Vision Workshops Approximate Cross Channel Color Mapping from Sparse Color Correspondences Hasan Sheikh Faridul1,2, Jurgen Stauder1, Jonathan Kervec1, Alain Tremeau´ 2 1Technicolor Research & Innovation, France 2University of Saint Etienne, CNRS, UMR 5516, France {hasan.sheikh-faridul,jurgen.stauder}@technicolor.com [email protected] Abstract change (second step of color mapping), mostly channel- wise [2, 6, 9, 13, 27–31] color mapping models are used. We propose a color mapping method that compensates Channel-wise means that color mapping models are inde- color differences between images having a common seman- pendently estimated for each color channels (such as red, tic content such as multiple views of a scene taken from green and blue). In a channel-wise model, to estimate a different viewpoints. A so-called color mapping model is channel of one view, contribution is taken only from the usually estimated from color correspondences selected from corresponding channel of the other view, that is: those images. In this work, we introduce a color map- = ( ) ping that model color change in two steps: first, nonlin- Rview2 fR Rview1 ear, channel-wise mapping; second, linear, cross-channel Gview2 = fG(Gview1) (1) mapping. Additionally, unlike many state of the art meth- Bview2 = fB(Bview1). ods, we estimate the model from sparse matches and do not require dense geometric correspondences. We show that Channel-wise methods can be further categorized into linear well known cross-channel color change can be estimated [2, 28, 30, 31] or nonlinear [6, 9, 12, 14, 21] models. from sparse color correspondence.
    [Show full text]
  • Color Appearance Models Second Edition
    Color Appearance Models Second Edition Mark D. Fairchild Munsell Color Science Laboratory Rochester Institute of Technology, USA Color Appearance Models Wiley–IS&T Series in Imaging Science and Technology Series Editor: Michael A. Kriss Formerly of the Eastman Kodak Research Laboratories and the University of Rochester The Reproduction of Colour (6th Edition) R. W. G. Hunt Color Appearance Models (2nd Edition) Mark D. Fairchild Published in Association with the Society for Imaging Science and Technology Color Appearance Models Second Edition Mark D. Fairchild Munsell Color Science Laboratory Rochester Institute of Technology, USA Copyright © 2005 John Wiley & Sons Ltd, The Atrium, Southern Gate, Chichester, West Sussex PO19 8SQ, England Telephone (+44) 1243 779777 This book was previously publisher by Pearson Education, Inc Email (for orders and customer service enquiries): [email protected] Visit our Home Page on www.wileyeurope.com or www.wiley.com All Rights Reserved. No part of this publication may be reproduced, stored in a retrieval system or transmitted in any form or by any means, electronic, mechanical, photocopying, recording, scanning or otherwise, except under the terms of the Copyright, Designs and Patents Act 1988 or under the terms of a licence issued by the Copyright Licensing Agency Ltd, 90 Tottenham Court Road, London W1T 4LP, UK, without the permission in writing of the Publisher. Requests to the Publisher should be addressed to the Permissions Department, John Wiley & Sons Ltd, The Atrium, Southern Gate, Chichester, West Sussex PO19 8SQ, England, or emailed to [email protected], or faxed to (+44) 1243 770571. This publication is designed to offer Authors the opportunity to publish accurate and authoritative information in regard to the subject matter covered.
    [Show full text]
  • 2019 Color and Imaging Conference Final Program and Abstracts
    CIC27 FINAL PROGRAM AND PROCEEDINGS Twenty-seventh Color and Imaging Conference Color Science and Engineering Systems, Technologies, and Applications October 21–25, 2019 Paris, France #CIC27 MCS 2019: 20th International Symposium on Multispectral Colour Science Sponsored by the Society for Imaging Science and Technology The papers in this volume represent the program of the Twenty-seventh Color and Imaging Conference (CIC27) held October 21-25, 2019, in Paris, France. Copyright 2019. IS&T: Society for Imaging Science and Technology 7003 Kilworth Lane Springfield, VA 22151 USA 703/642-9090; 703/642-9094 fax [email protected] www.imaging.org All rights reserved. The abstract book with full proceeding on flash drive, or parts thereof, may not be reproduced in any form without the written permission of the Society. ISBN Abstract Book: 978-0-89208-343-5 ISBN USB Stick: 978-0-89208-344-2 ISNN Print: 2166-9635 ISSN Online: 2169-2629 Contributions are reproduced from copy submitted by authors; no editorial changes have been made. Printed in France. Bienvenue à CIC27! Thank you for joining us in Paris at the Twenty-seventh Color and Imaging Conference CONFERENCE EXHIBITOR (CIC27), returning this year to Europe for an exciting week of courses, workshops, keynotes, AND SPONSOR and oral and poster presentations from all sides of the lively color and imaging community. I’ve attended every CIC since CIC13 in 2005, first coming to the conference as a young PhD student and monitoring nearly every course CIC had to offer. I have learned so much through- out the years and am now truly humbled to serve as the general chair and excited to welcome you to my hometown! The conference begins with an intensive one-day introduction to color science course, along with a full-day material appearance workshop organized by GDR APPAMAT, focusing on the different disciplinary fields around the appearance of materials, surfaces, and objects and how to measure, model, and reproduce them.
    [Show full text]
  • Color-Gamut Mapping in the Non-Uniform CIE-1931 Space with Perceptual Hue Fidelity Constraints for SMPTE ST.2094-40 Standard Chang Su, Li Tao and Yeong Taeg Kim
    SIP (2020), vol. 9, e12, page 1 of 14 © The Author(s), 2020. Published by Cambridge University Press in association with Asia Pacific Signal and Information Processing Association. This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited. doi:10.1017/ATSIP.2020.11 industrial technology advances Color-gamut mapping in the non-uniform CIE-1931 space with perceptual hue fidelity constraints for SMPTE ST.2094-40 standard chang su, li tao and yeong taeg kim As high-dynamic range (HDR) and wide-color gamut (WCG) contents become more and more popular in multimedia markets, color mapping of the distributed contents to different rendering devices plays a pivotal role in HDR distribution eco-systems. The widely used and economic gamut-clipping (GC)-based techniques perform poorly in mapping WCG contents to narrow gamut devices; and high-performance color-appearance model (CAM)-based techniques are computationally expensive to commercial applications. In this paper, we propose a novel color gamut mapping (CGM) algorithm to solve the problem. By introducing a color transition/protection zone (TPZ) and a set of perceptual hue fidelity constraints into the CIE-1931 space, the proposed algorithm directly carries out CGM in the perceptually non-uniform space, thus greatly decreases the computational complexity. The proposed TPZ effectively achieves a reasonable compromise between saturation preserving and details protection in out- of-gamut colors. The proposed hue fidelity constraints reference the measurements of human subjects’ visual responses, thus effectively preserve the perceptual hue of the original colors.
    [Show full text]
  • Spot Color Mapping in Flexi
    SA International Spot Color Mapping in Flexi CHAPTER 1│Spot Color Mapping in Flexi 1.1. About Spot Color Mapping There are 2 different spot color mapping settings in Flexi. One deals with the input values of spot colors and how they will be interpreted. The other deals with the output values of spot colors and how they will be printed. 1.1.1. Global Spot Color Mapping Global Spot Color mapping handles the input values of spot colors. Spot colors are actually completely name based. Because they are intended to be handled by custom spot color mapping in RIP software, it is the name that is the most crucial part of a spot color. For display purposes, design software often assign "alternative values" either in RGB or CMYK. RIP software uses these alternative values as a starting point to print the spot color. From that starting point, the output can be further fine tuned in Custom Spot Color Mapping. However, these alternative values are not always correct. Only design software that has a license for certain color libraries like Pantone TM know the correct values for the spot colors. With Global Spot Color Mapping, Flexi replaces the alternative values of incoming files with the correct values supplied by the color library man- ufacturers. 1.1.2. Custom Spot Color Mapping Custom Spot Color Mapping allows you to map spot colors to exact output values for your specific output device. Mapped colors will always print out using the output values set in the Custom Spot Color Mapping module, overriding any other color management settings.
    [Show full text]