Hans Hass' Under the Caribbean (1954, LI, AT, DE) was one of the world's first underwater colour films. As such, it provides a unique case study and raises interesting questions about the film's colour technology, combining 35 mm chromogenic negative and 16 mm Kodachrome processes with Technicolor imbibition printing in an interweaving of colour processes. Research into the vast amount of Hass' film material held at the Filmarchiv Austria has not yet revealed any of the original Kodachrome footage of this film nor its opticals. However, based on archival documents, it was possible to confirm and reconstruct the workflow Technicolor adopted for this film. Investigating the production history of Under the Caribbean not only provides film historical knowledge of this specific film, but also film technical insights into the production of other films of the early 50s, that also combine several colour processes. This research will be presented together with a discussion of the restoration possibilities offered by the source material, that is, the cut negative and several release prints.
{"title":"The colour Technology of Under the Caribbean (Hans Hass, 1954) Through a Comparison of Original Film Sources and Archival Documents","authors":"Rita Clemens","doi":"10.1002/col.22974","DOIUrl":"https://doi.org/10.1002/col.22974","url":null,"abstract":"<div>\u0000 \u0000 <p>Hans Hass' <i>Under the Caribbean</i> (1954, LI, AT, DE) was one of the world's first underwater colour films. As such, it provides a unique case study and raises interesting questions about the film's colour technology, combining 35 mm chromogenic negative and 16 mm Kodachrome processes with Technicolor imbibition printing in an interweaving of colour processes. Research into the vast amount of Hass' film material held at the Filmarchiv Austria has not yet revealed any of the original Kodachrome footage of this film nor its opticals. However, based on archival documents, it was possible to confirm and reconstruct the workflow Technicolor adopted for this film. Investigating the production history of <i>Under the Caribbean</i> not only provides film historical knowledge of this specific film, but also film technical insights into the production of other films of the early 50s, that also combine several colour processes. This research will be presented together with a discussion of the restoration possibilities offered by the source material, that is, the cut negative and several release prints.</p>\u0000 </div>","PeriodicalId":10459,"journal":{"name":"Color Research and Application","volume":"50 3","pages":"276-282"},"PeriodicalIF":1.2,"publicationDate":"2024-12-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143846153","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Skin color constancy under nonuniform correlated color temperatures (CCT) and multiple light sources has always been a hot issue in color science. A more high-quality skin color reproduction method has broad application prospects in camera photography, face recognition, and other fields. The processing process from the 14bit or 16bit RAW pictures taken by the camera to the final output of 8bit JPG pictures is called the image processing pipeline, in which the steps of the auto-white balance algorithm have a decisive impact on the skin color reproduction result. The traditional automatic white balance algorithm is based on hypothetical statistics. Moreover, the estimated illuminant color is obtained through illuminant estimation. However, the traditional grayscale world, perfect reflector, and other auto-white balance algorithms perform unsatisfactorily under non-uniform or complex light sources. The method based on sample statistics proposes a new solution to this problem from another aspect. The deep learning algorithm, especially the generative adversarial network (GAN) algorithm, is very suitable for establishing the mapping between pictures and has an excellent performance in the fields of image reconstruction, image translation, defogging, and coloring. This paper proposes a new solution to this problem. The asymmetric UNet3+ shape generator integrates better global and local information to obtain a more refined correction matrix incorporating details of the whole image. The discriminator is Patch-discriminator, which focuses more on image details by changing the attention field. The dataset used in this article is the Liverpool-Leeds Skin-color Database (LLSD) and some supplementary images, including the skin color of more than 960 subjects under D65 and different light sources. Finally, we calculate the CIEDE2000 color difference and some other image quality index between the test skin color JPEG picture corrected by the auto-white balance algorithm and the skin color under the corresponding D65 to evaluate the effect of white balance correction. The results show that the asymmetric GAN algorithm proposed in this paper can bring higher quality skin color reproduction results than the traditional auto-white balance algorithm and existing deep learning WB algorithm.
{"title":"Auto-White Balance Algorithm of Skin Color Based on Asymmetric Generative Adversarial Network","authors":"Sicong Zhou, Hesong Li, Wenjun Sun, Fanyi Zhou, Kaida Xiao","doi":"10.1002/col.22970","DOIUrl":"https://doi.org/10.1002/col.22970","url":null,"abstract":"<p>Skin color constancy under nonuniform correlated color temperatures (CCT) and multiple light sources has always been a hot issue in color science. A more high-quality skin color reproduction method has broad application prospects in camera photography, face recognition, and other fields. The processing process from the 14bit or 16bit RAW pictures taken by the camera to the final output of 8bit JPG pictures is called the image processing pipeline, in which the steps of the auto-white balance algorithm have a decisive impact on the skin color reproduction result. The traditional automatic white balance algorithm is based on hypothetical statistics. Moreover, the estimated illuminant color is obtained through illuminant estimation. However, the traditional grayscale world, perfect reflector, and other auto-white balance algorithms perform unsatisfactorily under non-uniform or complex light sources. The method based on sample statistics proposes a new solution to this problem from another aspect. The deep learning algorithm, especially the generative adversarial network (GAN) algorithm, is very suitable for establishing the mapping between pictures and has an excellent performance in the fields of image reconstruction, image translation, defogging, and coloring. This paper proposes a new solution to this problem. The asymmetric UNet3+ shape generator integrates better global and local information to obtain a more refined correction matrix incorporating details of the whole image. The discriminator is Patch-discriminator, which focuses more on image details by changing the attention field. The dataset used in this article is the Liverpool-Leeds Skin-color Database (LLSD) and some supplementary images, including the skin color of more than 960 subjects under D65 and different light sources. Finally, we calculate the CIEDE2000 color difference and some other image quality index between the test skin color JPEG picture corrected by the auto-white balance algorithm and the skin color under the corresponding D65 to evaluate the effect of white balance correction. The results show that the asymmetric GAN algorithm proposed in this paper can bring higher quality skin color reproduction results than the traditional auto-white balance algorithm and existing deep learning WB algorithm.</p>","PeriodicalId":10459,"journal":{"name":"Color Research and Application","volume":"50 3","pages":"266-275"},"PeriodicalIF":1.2,"publicationDate":"2024-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/col.22970","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143846076","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Color palette is a critical component of art, design, and lots of applications, providing the basis for organizing and utilizing colors to achieve specific objectives. However, generating color palettes from digital images presents unique challenges due to the complexity of colors in images. This review comprehensively investigates various techniques for generating color palettes from digital images and provides a thorough classification and discussion of these techniques from multiple perspectives. A color space must be selected to generate a color palette, and a generation method must be employed. This paper offers a concise overview of color spaces, an introduction to current palette generation methods, and an analysis of the metrics used to evaluate color differences between palettes. The review encompasses traditional manual methods and computer-aided automation methods, further categorized as histogram-based, clustering-based, and neural network-based methods. Discussion on the strengths, weaknesses, and applicability of existing methods are presented, and also opportunities for future research to enhance color palette generation from digital images are identified.
{"title":"Color Palette Generation From Digital Images: A Review","authors":"Yafan Gao, Jinxing Liang, Jie Yang","doi":"10.1002/col.22975","DOIUrl":"https://doi.org/10.1002/col.22975","url":null,"abstract":"<p>Color palette is a critical component of art, design, and lots of applications, providing the basis for organizing and utilizing colors to achieve specific objectives. However, generating color palettes from digital images presents unique challenges due to the complexity of colors in images. This review comprehensively investigates various techniques for generating color palettes from digital images and provides a thorough classification and discussion of these techniques from multiple perspectives. A color space must be selected to generate a color palette, and a generation method must be employed. This paper offers a concise overview of color spaces, an introduction to current palette generation methods, and an analysis of the metrics used to evaluate color differences between palettes. The review encompasses traditional manual methods and computer-aided automation methods, further categorized as histogram-based, clustering-based, and neural network-based methods. Discussion on the strengths, weaknesses, and applicability of existing methods are presented, and also opportunities for future research to enhance color palette generation from digital images are identified.</p>","PeriodicalId":10459,"journal":{"name":"Color Research and Application","volume":"50 3","pages":"250-265"},"PeriodicalIF":1.2,"publicationDate":"2024-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/col.22975","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143846092","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This study investigated how color variations in facial expressions influence our perception of emotions and health. Participants viewed color-manipulated (CIE LAB color space) face images depicting seven emotional states and indicated their perceptions of each image's emotion and health. Our results suggest that facial color influences the perception of threat-related emotions such as anger and disgust, as well as health perception. Increasing facial redness intensified the perception of anger, while increasing yellowness and lightness heightened the perception of disgust. Lightness affected perceptions of happiness and sadness, with lighter happy faces appearing happier and lighter sad faces appearing sadder. Additionally, enhancing redness and yellowness on faces led participants to perceive them as healthier. Our findings add to the existing literature and provide important insights into the role of colors in perceiving different emotions and health. These insights may significantly impact social interaction and communication, especially in situations where facial expressions play a critical role.
{"title":"Shades of Feeling: How Facial Color Variations Influence Emotional and Health Perception","authors":"Faeze Heydari, Majid Khalili-Ardali, Ali Yoonessi","doi":"10.1002/col.22968","DOIUrl":"https://doi.org/10.1002/col.22968","url":null,"abstract":"<div>\u0000 \u0000 <p>This study investigated how color variations in facial expressions influence our perception of emotions and health. Participants viewed color-manipulated (CIE LAB color space) face images depicting seven emotional states and indicated their perceptions of each image's emotion and health. Our results suggest that facial color influences the perception of threat-related emotions such as anger and disgust, as well as health perception. Increasing facial redness intensified the perception of anger, while increasing yellowness and lightness heightened the perception of disgust. Lightness affected perceptions of happiness and sadness, with lighter happy faces appearing happier and lighter sad faces appearing sadder. Additionally, enhancing redness and yellowness on faces led participants to perceive them as healthier. Our findings add to the existing literature and provide important insights into the role of colors in perceiving different emotions and health. These insights may significantly impact social interaction and communication, especially in situations where facial expressions play a critical role.</p>\u0000 </div>","PeriodicalId":10459,"journal":{"name":"Color Research and Application","volume":"50 3","pages":"240-249"},"PeriodicalIF":1.2,"publicationDate":"2024-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143845821","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Two-dimensional (2D) color appearance scales, including attributes such as whiteness, blackness, vividness, and depth, are closely aligned with our daily experiences and are thus highly sought after. This article presents a psychophysical experiment utilizing the magnitude estimation method under four levels of luminance (10, 100, 1000, and 10 000 cd/m2). Observers evaluated both traditional one-dimensional scales (brightness, colorfulness, and hue) and 2D scales (whiteness, blackness, vividness, and depth). The collected data were utilized to evaluate existing color appearance models, including CIECAM16, CAM16-UCS, V*ab, and D*ab. Subsequently, new scales derived from the CAM16-UCS model were developed to more accurately reflect the experimental data. These scales were further validated against independent data sets. The findings indicated that the scales developed from this study outperformed existing models when applied to external data sets. Moreover, Berns's scales for vividness and depth demonstrated strong predictive accuracy. Finally, four simple and accurate 2D scales are proposed.
{"title":"Development of Vividness, Blackness, Depth, and Whiteness Scales Under High Range of Luminance Levels","authors":"Molin Li, Yuechen Zhu, Ming Ronnier Luo","doi":"10.1002/col.22966","DOIUrl":"https://doi.org/10.1002/col.22966","url":null,"abstract":"<div>\u0000 \u0000 <p>Two-dimensional (2D) color appearance scales, including attributes such as whiteness, blackness, vividness, and depth, are closely aligned with our daily experiences and are thus highly sought after. This article presents a psychophysical experiment utilizing the magnitude estimation method under four levels of luminance (10, 100, 1000, and 10 000 cd/m<sup>2</sup>). Observers evaluated both traditional one-dimensional scales (brightness, colorfulness, and hue) and 2D scales (whiteness, blackness, vividness, and depth). The collected data were utilized to evaluate existing color appearance models, including CIECAM16, CAM16-UCS, <i>V</i>\u0000 <sup>\u0000 <i>*</i>\u0000 </sup>\u0000 <sub>\u0000 <i>ab</i>\u0000 </sub>, and <i>D</i>\u0000 <sup>\u0000 <i>*</i>\u0000 </sup>\u0000 <sub>\u0000 <i>ab</i>\u0000 </sub>. Subsequently, new scales derived from the CAM16-UCS model were developed to more accurately reflect the experimental data. These scales were further validated against independent data sets. The findings indicated that the scales developed from this study outperformed existing models when applied to external data sets. Moreover, Berns's scales for vividness and depth demonstrated strong predictive accuracy. Finally, four simple and accurate 2D scales are proposed.</p>\u0000 </div>","PeriodicalId":10459,"journal":{"name":"Color Research and Application","volume":"50 3","pages":"221-239"},"PeriodicalIF":1.2,"publicationDate":"2024-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143846159","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Color appearance characterization of stimuli is important in color science, with great efforts made to develop uniform color spaces (UCSs) and color appearance models (CAMs). These UCSs and CAMs, however, were mainly developed for stimuli with luminance below the diffuse white in a standard dynamic range (SDR) scene. This study was carefully designed to investigate the color appearance of self-luminous highlight stimuli, in terms of relative brightness, in a high dynamic range (HDR) scene, and to see whether a new model for characterizing the brightness of highlights is needed. The observer viewed a dark stimulus (Stimulus 0) and a self-luminous highlight (Stimulus 2) simultaneously under a certain diffuse white luminance level, which was an HDR scene, and adjusted the luminance of the other self-luminous highlight (Stimulus 1) so that its brightness appeared in the middle of Stimuli 0 and 2. The luminance of Stimulus 2 was designed to as high as 49 000 cd/m2, the luminance contrast between Stimulus 2 and Stimulus 0 was as high as 72 045:1, and the diffuse white luminance level was designed to between 10 and 11 000 cd/m2. The results suggested that the UCSs and CAMs adopting a power function to characterize the non-linear relationship between the relative luminance and perceived lightness (e.g., IPT, CIECAM02, and CIELAB) had much better performance to characterize the relative relationship of the lightness among the stimuli. Significant variations, however, were found among these UCSs and CAMs in characterizing the absolute magnitude of lightness and chroma, which merits further investigations.
{"title":"Characterization of Color Appearance of Self-Luminous Highlights Under HDR Scenes, Part I: Brightness for Neutral Stimuli","authors":"Hongbing Wang, Minchen Wei","doi":"10.1002/col.22965","DOIUrl":"https://doi.org/10.1002/col.22965","url":null,"abstract":"<div>\u0000 \u0000 <p>Color appearance characterization of stimuli is important in color science, with great efforts made to develop uniform color spaces (UCSs) and color appearance models (CAMs). These UCSs and CAMs, however, were mainly developed for stimuli with luminance below the diffuse white in a standard dynamic range (SDR) scene. This study was carefully designed to investigate the color appearance of self-luminous highlight stimuli, in terms of relative brightness, in a high dynamic range (HDR) scene, and to see whether a new model for characterizing the brightness of highlights is needed. The observer viewed a dark stimulus (Stimulus 0) and a self-luminous highlight (Stimulus 2) simultaneously under a certain diffuse white luminance level, which was an HDR scene, and adjusted the luminance of the other self-luminous highlight (Stimulus 1) so that its brightness appeared in the middle of Stimuli 0 and 2. The luminance of Stimulus 2 was designed to as high as 49 000 cd/m<sup>2</sup>, the luminance contrast between Stimulus 2 and Stimulus 0 was as high as 72 045:1, and the diffuse white luminance level was designed to between 10 and 11 000 cd/m<sup>2</sup>. The results suggested that the UCSs and CAMs adopting a power function to characterize the non-linear relationship between the relative luminance and perceived lightness (e.g., IPT, CIECAM02, and CIELAB) had much better performance to characterize the relative relationship of the lightness among the stimuli. Significant variations, however, were found among these UCSs and CAMs in characterizing the absolute magnitude of lightness and chroma, which merits further investigations.</p>\u0000 </div>","PeriodicalId":10459,"journal":{"name":"Color Research and Application","volume":"50 3","pages":"212-220"},"PeriodicalIF":1.2,"publicationDate":"2024-11-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143845957","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Individual differences are a prominent feature of normal color vision and range from variations in sensitivity to perception and color naming. Corrections for differences in spectral sensitivity are common, and there is growing interest in calibrating displays for the sensitivity of an individual observer. In contrast, few studies have explored calibrations for aspects of color appearance. We developed a technique for adjusting images based directly on an individual's hue percepts and illustrate the principle of the approach using a set of hue scaling functions measured previously for a large sample of color-normal observers (Emery et al. PNAS 2023). Colors in an image are mapped onto the average scaling function to define the hue perceived by the average “standard observer.” This hue is then mapped back to the chromaticity that would elicit the same response in any individual observer. With this correction, different observers – each looking at physically different images calibrated for their own hue percepts – should in principle agree on the perceived colors. Adjustments of this kind could be easily implemented on standard displays, because they require only measures of hue percepts and not spectral sensitivity, and could lead to greater consistency in the perception and communication about color across individuals with potentially widely different perceptual experiences of color.
{"title":"Correcting images for individual differences in color appearance","authors":"Camilla Simoncelli, Michael A. Webster","doi":"10.1002/col.22963","DOIUrl":"https://doi.org/10.1002/col.22963","url":null,"abstract":"<p>Individual differences are a prominent feature of normal color vision and range from variations in sensitivity to perception and color naming. Corrections for differences in spectral sensitivity are common, and there is growing interest in calibrating displays for the sensitivity of an individual observer. In contrast, few studies have explored calibrations for aspects of color appearance. We developed a technique for adjusting images based directly on an individual's hue percepts and illustrate the principle of the approach using a set of hue scaling functions measured previously for a large sample of color-normal observers (Emery et al. PNAS 2023). Colors in an image are mapped onto the average scaling function to define the hue perceived by the average “standard observer.” This hue is then mapped back to the chromaticity that would elicit the same response in any individual observer. With this correction, different observers – each looking at physically different images calibrated for their own hue percepts – should in principle agree on the perceived colors. Adjustments of this kind could be easily implemented on standard displays, because they require only measures of hue percepts and not spectral sensitivity, and could lead to greater consistency in the perception and communication about color across individuals with potentially widely different perceptual experiences of color.</p>","PeriodicalId":10459,"journal":{"name":"Color Research and Application","volume":"50 2","pages":"172-186"},"PeriodicalIF":1.2,"publicationDate":"2024-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/col.22963","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143455893","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yuan Liu, Yonggang Tian, Yuyi Liu, Keran Cao, Chuanyin Liu
Garden color plays a key role in creating a spatial atmosphere, artistic conception and regional landscape characteristics. The Chinese classical garden has a profound traditional Chinese culture at its core, and its color collocation is a microcosm of Chinese culture. An in-depth exploration of the color matching of classical gardens is helpful for promoting the organic combination of modern urban construction needs and traditional culture. Therefore, this paper takes Tang-style gardens in Xi'an as an example and selects five typical classical gardens to study the color characteristics of garden elements, uses natural color system (NCS) to record color data from classical garden elements, establishes a color map, and quantitatively evaluates the color harmony of classical garden elements on the basis of Moon Spencer (M-S) color harmony theory. Research has revealed that the color system of Tang-style gardens in Xi'an is unique and includes 244 NCS standard colors with a wide distribution of hues. The study distinguishes between humanistic environment colors and natural environment colors that vary greatly with season, and provides a wide color range that reflects the beauty of nature. The color range of the humanistic environment appears calm and solemn, with colors having low blackness and low chromaticness in general. There is a certain difference in color harmony between natural and cultural environment colors, but the overall color harmony is good. This paper not only enriches the color theory system of Tang-style gardens but also provides a reference for promoting the scientific application of traditional urban colors in modern urban landscape construction.
{"title":"Color collection and data analysis of classical gardens—A case study of Tang style gardens in Xi'an","authors":"Yuan Liu, Yonggang Tian, Yuyi Liu, Keran Cao, Chuanyin Liu","doi":"10.1002/col.22961","DOIUrl":"https://doi.org/10.1002/col.22961","url":null,"abstract":"<p>Garden color plays a key role in creating a spatial atmosphere, artistic conception and regional landscape characteristics. The Chinese classical garden has a profound traditional Chinese culture at its core, and its color collocation is a microcosm of Chinese culture. An in-depth exploration of the color matching of classical gardens is helpful for promoting the organic combination of modern urban construction needs and traditional culture. Therefore, this paper takes Tang-style gardens in Xi'an as an example and selects five typical classical gardens to study the color characteristics of garden elements, uses natural color system (NCS) to record color data from classical garden elements, establishes a color map, and quantitatively evaluates the color harmony of classical garden elements on the basis of Moon Spencer (M-S) color harmony theory. Research has revealed that the color system of Tang-style gardens in Xi'an is unique and includes 244 NCS standard colors with a wide distribution of hues. The study distinguishes between humanistic environment colors and natural environment colors that vary greatly with season, and provides a wide color range that reflects the beauty of nature. The color range of the humanistic environment appears calm and solemn, with colors having low blackness and low chromaticness in general. There is a certain difference in color harmony between natural and cultural environment colors, but the overall color harmony is good. This paper not only enriches the color theory system of Tang-style gardens but also provides a reference for promoting the scientific application of traditional urban colors in modern urban landscape construction.</p>","PeriodicalId":10459,"journal":{"name":"Color Research and Application","volume":"50 3","pages":"190-211"},"PeriodicalIF":1.2,"publicationDate":"2024-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143845884","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Just-noticeable color difference (JNCD) is important in color specification and characterization. The commonly referenced specification of JNCD (i.e., 0.004 or 0.0033 u′v′ unit) is thought to originate from the MacAdam ellipses, which were derived using 2° color stimuli and characterized using the CIE 1931 2° color matching functions (CMFs). However, there is no universally agreed or clear definition of JNCD. Also, such a specification is widely used in various ways based on an assumption that it is applicable regardless of the actual size of the stimuli and CMFs. In this study, an experiment using a constant stimuli method was carried out. The human observers evaluated a series pairs of test and reference stimuli, with a field of view (FOV) of 2° or 10°. The chromaticities of the test stimuli were carefully calibrated using four standard CIE CMFs (i.e., CIE 1931 2°, 1964 10°, 2006 2°, and 10° CMFs). The results suggested that the widely used specification of JNCD seems to be derived based on the one standard deviation ellipses, the use of these four CMFs has little effect on the specification, and the JNCD value for stimuli with an FOV of 10° is 0.0025–0.0027 u′v′ unit depending on the CMFs.
{"title":"Specification of just-noticeable color difference for 2° and 10° stimuli using different color matching functions","authors":"Mengjing Zhao, Minchen Wei","doi":"10.1002/col.22962","DOIUrl":"https://doi.org/10.1002/col.22962","url":null,"abstract":"<p>Just-noticeable color difference (JNCD) is important in color specification and characterization. The commonly referenced specification of JNCD (i.e., 0.004 or 0.0033 <i>u</i>′<i>v</i>′ unit) is thought to originate from the MacAdam ellipses, which were derived using 2° color stimuli and characterized using the CIE 1931 2° color matching functions (CMFs). However, there is no universally agreed or clear definition of JNCD. Also, such a specification is widely used in various ways based on an assumption that it is applicable regardless of the actual size of the stimuli and CMFs. In this study, an experiment using a constant stimuli method was carried out. The human observers evaluated a series pairs of test and reference stimuli, with a field of view (FOV) of 2° or 10°. The chromaticities of the test stimuli were carefully calibrated using four standard CIE CMFs (i.e., CIE 1931 2°, 1964 10°, 2006 2°, and 10° CMFs). The results suggested that the widely used specification of JNCD seems to be derived based on the one standard deviation ellipses, the use of these four CMFs has little effect on the specification, and the JNCD value for stimuli with an FOV of 10° is 0.0025–0.0027 <i>u</i>′<i>v</i>′ unit depending on the CMFs.</p>","PeriodicalId":10459,"journal":{"name":"Color Research and Application","volume":"50 2","pages":"161-171"},"PeriodicalIF":1.2,"publicationDate":"2024-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/col.22962","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143455838","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In 2022, CIE recommend the CIECAM16 color appearance model to replace the current CIECAM02 model for color management systems. In this paper we will report on how CIECAM16 was developed, describe the differences between CIECAM16 and CIECAM02, describe the phenomena CIECAM16 can predict, and show the performance of CIECAM16 in the prediction of perceptual color attributes. We will then discuss the numerical determination of the domain and range of the CIECAM16 forward transformation. Finally, the domain and ranges in various conditions are visualized. CIECAM16 is capable of the accurate prediction of color appearance under a wide range of viewing conditions and with the domain and range problems now solved, it can be effectively applied to cross-media color image reproduction. It can also be used to estimate the color rendering properties of light sources and for the establishment of a uniform color space for color difference evaluation.
{"title":"The development of the CIECAM16 and visualization of its domain and range","authors":"Cheng Gao, Kaida Xiao, Mike Pointer, Changjun Li","doi":"10.1002/col.22959","DOIUrl":"https://doi.org/10.1002/col.22959","url":null,"abstract":"<p>In 2022, CIE recommend the CIECAM16 color appearance model to replace the current CIECAM02 model for color management systems. In this paper we will report on how CIECAM16 was developed, describe the differences between CIECAM16 and CIECAM02, describe the phenomena CIECAM16 can predict, and show the performance of CIECAM16 in the prediction of perceptual color attributes. We will then discuss the numerical determination of the domain and range of the CIECAM16 forward transformation. Finally, the domain and ranges in various conditions are visualized. CIECAM16 is capable of the accurate prediction of color appearance under a wide range of viewing conditions and with the domain and range problems now solved, it can be effectively applied to cross-media color image reproduction. It can also be used to estimate the color rendering properties of light sources and for the establishment of a uniform color space for color difference evaluation.</p>","PeriodicalId":10459,"journal":{"name":"Color Research and Application","volume":"50 2","pages":"144-160"},"PeriodicalIF":1.2,"publicationDate":"2024-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/col.22959","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143456000","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}