首页 > 最新文献

Displays最新文献

英文 中文
Visible light-activated ultra-long orange afterglow phosphor Sr1−x Ga2O4:xCu2+
IF 3.7 2区 工程技术 Q1 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-03-15 DOI: 10.1016/j.displa.2025.103028
Li Hao, Gongchun Zhang, Qingping Wang, Ling Su, Lijuan Feng, Guiquan Jiang
Because of both their independence from external light sources and their energy-saving properties, efficient visible light-activated long persistent luminescence (LPL) phosphors are in high demand. Designing and developing such materials, however, remains a significant challenge. This report describes new LPL phosphors doped with transition metal ions (Sr1−x Ga2O4:xCu2+), which address the challenge of excitation of red light-emitting LPL phosphors by visible light. Visible light excitation is achieved by adjusting the calcination atmosphere during preparation of the phosphor. When the Cu2+ ion doping level is 2 %, the phosphor can be excited by both UV and visible light. Upon excitation by visible light, the main emission peak occurs at approximately 610 nm; this is observed as orange light, with an afterglow duration of over 11 h. Investigation of the emission spectra after different delay times, together with the photoluminescence and thermoluminescence spectra, revealed a mechanism that involves trapping and de-trapping of electrons. The unique long afterglow and excitation properties, together with the emission characteristics of SrGa2O4:0.02Cu2+, are particularly suited for applications such as night time signage and dynamic anti-counterfeiting. This research provides new directions for the design of visible light-activated LPL phosphors.
{"title":"Visible light-activated ultra-long orange afterglow phosphor Sr1−x Ga2O4:xCu2+","authors":"Li Hao,&nbsp;Gongchun Zhang,&nbsp;Qingping Wang,&nbsp;Ling Su,&nbsp;Lijuan Feng,&nbsp;Guiquan Jiang","doi":"10.1016/j.displa.2025.103028","DOIUrl":"10.1016/j.displa.2025.103028","url":null,"abstract":"<div><div>Because of both their independence from external light sources and their energy-saving properties, efficient visible light-activated long persistent luminescence (LPL) phosphors are in high demand. Designing and developing such materials, however, remains a significant challenge. This report describes new LPL phosphors doped with transition metal ions (Sr<sub>1−x</sub> Ga<sub>2</sub>O<sub>4</sub>:xCu<sup>2+</sup>), which address the challenge of excitation of red light-emitting LPL phosphors by visible light. Visible light excitation is achieved by adjusting the calcination atmosphere during preparation of the phosphor. When the Cu<sup>2+</sup> ion doping level is 2 %, the phosphor can be excited by both UV and visible light. Upon excitation by visible light, the main emission peak occurs at approximately 610 nm; this is observed as orange light, with an afterglow duration of over 11 h. Investigation of the emission spectra after different delay times, together with the photoluminescence and thermoluminescence spectra, revealed a mechanism that involves trapping and de-trapping of electrons. The unique long afterglow and excitation properties, together with the emission characteristics of SrGa<sub>2</sub>O<sub>4</sub>:0.02Cu<sup>2+</sup>, are particularly suited for applications such as night time signage and dynamic anti-counterfeiting. This research provides new directions for the design of visible light-activated LPL phosphors.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 103028"},"PeriodicalIF":3.7,"publicationDate":"2025-03-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143644978","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Quantifying the impact of night-shift display modes on evening melatonin production
IF 3.7 2区 工程技术 Q1 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-03-14 DOI: 10.1016/j.displa.2025.103027
Jiawen Mao , Yingying Huang , Mingyuan Zhao , Feng Zhang , Xin Li , Qi Dai
The use of self-emitting devices in the evening can significantly suppress melatonin secretion, leading to sleep onset delays and circadian disruption. To minimize these undesirable effects, various night-shift display modes have been developed. In this study, we conducted a human factor experiment to investigate the effectiveness of night-shift modes in maintaining evening melatonin production. The study adopted four different night-shift modes and a normal mode for smartphones, which were applied during the three hours before bedtime. The four different night-shift modes employed constant and dynamically reduced correlated color temperatures (CCTs) and luminance levels. Under modes with constant low luminance and CCTs (leading to low melanopic lux), participants had significantly higher melatonin production than those in the normal mode (with high melanopic lux). This result indicates that a reduction in melanopic lux is effective in mitigating the negative effect of display light on evening melatonin production. Regarding the dynamic modes, despite achieving lower melanopic lux levels before sleep, they exhibited relatively high melanopic lux during the early hours of the experiment. Thus, it is necessary to quantify the cumulative time effects of melanopic lux over various periods when evaluating its impact on melatonin production. It was found that the cumulative Melanopic Equivalent Daylight Illuminance (ʃ m-EDI dt) over two hours before bedtime could be an effective metric, which applies to both dynamic and static display modes. Based on the metric, a value below 35 lx·h in the evening is considered effective in reducing the attenuation of melatonin production.
{"title":"Quantifying the impact of night-shift display modes on evening melatonin production","authors":"Jiawen Mao ,&nbsp;Yingying Huang ,&nbsp;Mingyuan Zhao ,&nbsp;Feng Zhang ,&nbsp;Xin Li ,&nbsp;Qi Dai","doi":"10.1016/j.displa.2025.103027","DOIUrl":"10.1016/j.displa.2025.103027","url":null,"abstract":"<div><div>The use of self-emitting devices in the evening can significantly suppress melatonin secretion, leading to sleep onset delays and circadian disruption. To minimize these undesirable effects, various night-shift display modes have been developed. In this study, we conducted a human factor experiment to investigate the effectiveness of night-shift modes in maintaining evening melatonin production. The study adopted four different night-shift modes and a normal mode for smartphones, which were applied during the three hours before bedtime. The four different night-shift modes employed constant and dynamically reduced correlated color temperatures (CCTs) and luminance levels. Under modes with constant low luminance and CCTs (leading to low melanopic lux), participants had significantly higher melatonin production than those in the normal mode (with high melanopic lux). This result indicates that a reduction in melanopic lux is effective in mitigating the negative effect of display light on evening melatonin production. Regarding the dynamic modes, despite achieving lower melanopic lux levels before sleep, they exhibited relatively high melanopic lux during the early hours of the experiment. Thus, it is necessary to quantify the cumulative time effects of melanopic lux over various periods when evaluating its impact on melatonin production. It was found that the cumulative Melanopic Equivalent Daylight Illuminance (<em>ʃ</em> m-EDI <em>dt</em>) over two hours before bedtime could be an effective metric, which applies to both dynamic and static display modes. Based on the metric, a value below 35 lx·h in the evening is considered effective in reducing the attenuation of melatonin production.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 103027"},"PeriodicalIF":3.7,"publicationDate":"2025-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143644977","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The interaction efficiency of different visual areas on a virtual reality interaction space: Standing versus sitting posture
IF 3.7 2区 工程技术 Q1 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-03-13 DOI: 10.1016/j.displa.2025.103025
Jutao Li , Jingxuan Yuan , Yanqun Huang , Xiaotian Liang , Xueqin Huang , Jian Zhang
User posture and visual area can impact searching-selecting efficiency, thereby influencing interaction efficiency. Virtual Reality (VR) devices extend the traditional two-dimensional interface to a three-dimensional interaction space, requiring users to alter their body postures when using them. This study compared the searching-selecting efficiency of different postures (standing and sitting) and visual areas in a VR environment. An experiment measured thirty-five participants’ searching and selecting times for targets under different conditions. A two-way analysis of variance (ANOVA) indicated that the interaction efficiency was generally higher for participants in a sitting posture than in a standing posture. Participants spent less time searching for left side targets than in other visual areas. However, the interaction effect between posture and visual area was not significant. These results suggest that posture and visual areas respectively influence searching-selecting efficiency in VR environments. These findings could provide insights for VR developers to enhance interaction design, particularly interface design, for different user postures, thus supporting interaction efficiency and user experience.
{"title":"The interaction efficiency of different visual areas on a virtual reality interaction space: Standing versus sitting posture","authors":"Jutao Li ,&nbsp;Jingxuan Yuan ,&nbsp;Yanqun Huang ,&nbsp;Xiaotian Liang ,&nbsp;Xueqin Huang ,&nbsp;Jian Zhang","doi":"10.1016/j.displa.2025.103025","DOIUrl":"10.1016/j.displa.2025.103025","url":null,"abstract":"<div><div>User posture and visual area can impact searching-selecting efficiency, thereby influencing interaction efficiency. Virtual Reality (VR) devices extend the traditional two-dimensional interface to a three-dimensional interaction space, requiring users to alter their body postures when using them. This study compared the searching-selecting efficiency of different postures (standing and sitting) and visual areas in a VR environment. An experiment measured thirty-five participants’ searching and selecting times for targets under different conditions. A two-way analysis of variance (ANOVA) indicated that the interaction efficiency was generally higher for participants in a sitting posture than in a standing posture. Participants spent less time searching for left side targets than in other visual areas. However, the interaction effect between posture and visual area was not significant. These results suggest that posture and visual areas respectively influence searching-selecting efficiency in VR environments. These findings could provide insights for VR developers to enhance interaction design, particularly interface design, for different user postures, thus supporting interaction efficiency and user experience.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 103025"},"PeriodicalIF":3.7,"publicationDate":"2025-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143644976","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Low-Power integrated driver circuits with multiple outputs based on Thin-Film transistors for mobile micro Light-Emitting diode displays
IF 3.7 2区 工程技术 Q1 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-03-11 DOI: 10.1016/j.displa.2025.103022
Ye-Rim Jeong , Eun Kyo Jung , Kook Chul Moon , Hwarim Im , Yong-Sang Kim
This article proposes low-power integrated driver circuits based on low-temperature polycrystalline silicon (LTPS) thin-film transistors (TFTs) for small and medium-sized mobile micro light-emitting diode (μLED) displays. Two integrated driver circuits can output four signals to operate the μLED pixel circuit: the integrated scan driver outputs two scan signals, and the integrated emission driver outputs an emission signal and a sweep signal. Moreover, driver circuits are designed to enable the always-on-display (AOD) mode, which consumes less power than the normal mode. Compared to the normal mode, the proposed circuits do not output unnecessary signals in the AOD mode, reducing power consumption by 38.0 and 29.6% in the integrated scan and emission circuits. In addition, compared to the gate driver circuits with a single output signal. The proposed integrated scan and emission driver circuits reduce power consumption by 16.6 (25.2) and 42.5 (51.3)% in the normal (AOD) mode, respectively, by sharing the control unit for multiple outputs. We also fabricated the proposed driver circuits and verified that the proposed circuits successfully output multiple signals for mobile μLED pixel circuits.
{"title":"Low-Power integrated driver circuits with multiple outputs based on Thin-Film transistors for mobile micro Light-Emitting diode displays","authors":"Ye-Rim Jeong ,&nbsp;Eun Kyo Jung ,&nbsp;Kook Chul Moon ,&nbsp;Hwarim Im ,&nbsp;Yong-Sang Kim","doi":"10.1016/j.displa.2025.103022","DOIUrl":"10.1016/j.displa.2025.103022","url":null,"abstract":"<div><div>This article proposes low-power integrated driver circuits based on low-temperature polycrystalline silicon (LTPS) thin-film transistors (TFTs) for small and medium-sized mobile micro light-emitting diode (μLED) displays. Two integrated driver circuits can output four signals to operate the μLED pixel circuit: the integrated scan driver outputs two scan signals, and the integrated emission driver outputs an emission signal and a sweep signal. Moreover, driver circuits are designed to enable the always-on-display (AOD) mode, which consumes less power than the normal mode. Compared to the normal mode, the proposed circuits do not output unnecessary signals in the AOD mode, reducing power consumption by 38.0 and 29.6% in the integrated scan and emission circuits. In addition, compared to the gate driver circuits with a single output signal. The proposed integrated scan and emission driver circuits reduce power consumption by 16.6 (25.2) and 42.5 (51.3)% in the normal (AOD) mode, respectively, by sharing the control unit for multiple outputs. We also fabricated the proposed driver circuits and verified that the proposed circuits successfully output multiple signals for mobile μLED pixel circuits.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 103022"},"PeriodicalIF":3.7,"publicationDate":"2025-03-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143629179","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An AI-driven framework for perceived display spectra: The effects of dimming, observer age, and viewing distance
IF 3.7 2区 工程技术 Q1 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-03-10 DOI: 10.1016/j.displa.2025.103024
N. Senyer , D. Durmus
Displays have become ubiquitous in modern society, serving as pervasive light sources that exert visual and non-visual effects on human physiology and behavior. Despite their widespread use and impact, a universal framework for characterizing perceived display light output across various viewing conditions still needs to be developed. This study introduces a novel, AI-driven framework for comprehensive perceived display light output characterization, accounting for the effects of observer age, viewing distance, and display dimming. The framework employs a deep neural network (DNN) trained on an extensive dataset of measured display spectra to predict spectral power distributions (SPDs) from RGB inputs. To simulate real-world scenarios, the DNN-predicted SPDs were transformed to account for viewing distance (36 cm–71 cm), display dimming (0–100 %), and observer age (1–100 years). The initial model achieved high accuracy (R2avg = 0.99), maintaining robust performance even for challenging cases (R2 > 0.94). Results show high accuracy in predicting photometric, colorimetric, and circadian measures. Future research will incorporate other parameters to the proposed framework.
{"title":"An AI-driven framework for perceived display spectra: The effects of dimming, observer age, and viewing distance","authors":"N. Senyer ,&nbsp;D. Durmus","doi":"10.1016/j.displa.2025.103024","DOIUrl":"10.1016/j.displa.2025.103024","url":null,"abstract":"<div><div>Displays have become ubiquitous in modern society, serving as pervasive light sources that exert visual and non-visual effects on human physiology and behavior. Despite their widespread use and impact, a universal framework for characterizing perceived display light output across various viewing conditions still needs to be developed. This study introduces a novel, AI-driven framework for comprehensive perceived display light output characterization, accounting for the effects of observer age, viewing distance, and display dimming. The framework employs a deep neural network (DNN) trained on an extensive dataset of measured display spectra to predict spectral power distributions (SPDs) from RGB inputs. To simulate real-world scenarios, the DNN-predicted SPDs were transformed to account for viewing distance (36 cm–71 cm), display dimming (0–100 %), and observer age (1–100 years). The initial model achieved high accuracy (<em>R</em>2<sub>avg</sub> = 0.99), maintaining robust performance even for challenging cases (<em>R</em><sup>2</sup> &gt; 0.94). Results show high accuracy in predicting photometric, colorimetric, and circadian measures. Future research will incorporate other parameters to the proposed framework.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 103024"},"PeriodicalIF":3.7,"publicationDate":"2025-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143631880","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hedonic or Utilitarian: The effect of video and text styles on consumers’ visual attention, mental imagery quality, and purchase intention in online food ordering interfaces
IF 3.7 2区 工程技术 Q1 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-03-06 DOI: 10.1016/j.displa.2025.103016
Mengyao Qi , Kenta Ono , Lujin Mao , Makoto Watanabe , Jinghua Huang
Given the substantial resources allocated to video and text presentations, indiscriminately applying video and text presentations regardless of information style may not improve sales or enhance the consumers’ ordering experience, and can even lead to unnecessary resource waste. Therefore, this study aims to examine the effects of video and text styles (hedonic vs. utilitarian) on consumers’ visual attention and subjective assessments. The results indicate that the utilitarian text style could be a favorable choice for online food ordering interfaces, as it not only captures greater visual attention but also enhances users’ mental imagery quality and purchase intention. Additionally, the selection of video styles should be guided by the specific objectives of the context: the utilitarian video style appears to be a more effective option for encouraging purchasing behavior, as it is associated with significantly higher fixation counts, while the hedonic video style may be more conducive to enhancing user engagement and extending online duration, as indicated by its longer average fixation duration. These findings provide valuable insights into the effectiveness of short-form video and text across different information styles, contributing to the design of more user-friendly online food ordering interfaces.
{"title":"Hedonic or Utilitarian: The effect of video and text styles on consumers’ visual attention, mental imagery quality, and purchase intention in online food ordering interfaces","authors":"Mengyao Qi ,&nbsp;Kenta Ono ,&nbsp;Lujin Mao ,&nbsp;Makoto Watanabe ,&nbsp;Jinghua Huang","doi":"10.1016/j.displa.2025.103016","DOIUrl":"10.1016/j.displa.2025.103016","url":null,"abstract":"<div><div>Given the substantial resources allocated to video and text presentations, indiscriminately applying video and text presentations regardless of information style may not improve sales or enhance the consumers’ ordering experience, and can even lead to unnecessary resource waste. Therefore, this study aims to examine the effects of video and text styles (hedonic vs. utilitarian) on consumers’ visual attention and subjective assessments. The results indicate that the utilitarian text style could be a favorable choice for online food ordering interfaces, as it not only captures greater visual attention but also enhances users’ mental imagery quality and purchase intention. Additionally, the selection of video styles should be guided by the specific objectives of the context: the utilitarian video style appears to be a more effective option for encouraging purchasing behavior, as it is associated with significantly higher fixation counts, while the hedonic video style may be more conducive to enhancing user engagement and extending online duration, as indicated by its longer average fixation duration. These findings provide valuable insights into the effectiveness of short-form video and text across different information styles, contributing to the design of more user-friendly online food ordering interfaces.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 103016"},"PeriodicalIF":3.7,"publicationDate":"2025-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143591643","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Tailored information display: Effects of background colour and line spacing on visual search across different character types – An eye-tracking study
IF 3.7 2区 工程技术 Q1 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-03-04 DOI: 10.1016/j.displa.2025.103019
Xingkai Wang , Bingyuan Wang , Lina Xu , Luwen Yu
In the era of expanding digital information, optimizing display factors has become increasingly critical for human–computer interaction. However, the broader impacts of background colour and line spacing on interfaces with different character types, under controlled luminance contrast remain underexplored. This study investigates the effects of background colour and line spacing on search performance, eye-tracking metrics, and subjective reports across English, Chinese, and numerals in controlled luminance conditions using a pseudo-text visual search task. Findings reveal the distinct impacts and mechanisms for each character type. Red background significantly hindered search performance, particularly in English, while cyan improved accuracy and subjective ratings across all characters. Colour difference emerged as a critical determinant of visual comfort, and its influence is moderated by pupil diameter. Double line spacing provided an optimal balance between search performance and minimizing eye movement effort, further increases in line spacing impair search speed, especially in numerals. Fixation duration and fixation count mediated the effects of line spacing on legibility in numerals and English, respectively. These findings offer actionable insights for optimizing interfaces with different character types and highlight the need to explore further how display factors contribute to visual experience in complex scripts.
{"title":"Tailored information display: Effects of background colour and line spacing on visual search across different character types – An eye-tracking study","authors":"Xingkai Wang ,&nbsp;Bingyuan Wang ,&nbsp;Lina Xu ,&nbsp;Luwen Yu","doi":"10.1016/j.displa.2025.103019","DOIUrl":"10.1016/j.displa.2025.103019","url":null,"abstract":"<div><div>In the era of expanding digital information, optimizing display factors has become increasingly critical for human–computer interaction. However, the broader impacts of background colour and line spacing on interfaces with different character types, under controlled luminance contrast remain underexplored. This study investigates the effects of background colour and line spacing on search performance, eye-tracking metrics, and subjective reports across English, Chinese, and numerals in controlled luminance conditions using a pseudo-text visual search task. Findings reveal the distinct impacts and mechanisms for each character type. Red background significantly hindered search performance, particularly in English, while cyan improved accuracy and subjective ratings across all characters. Colour difference emerged as a critical determinant of visual comfort, and its influence is moderated by pupil diameter. Double line spacing provided an optimal balance between search performance and minimizing eye movement effort, further increases in line spacing impair search speed, especially in numerals. Fixation duration and fixation count mediated the effects of line spacing on legibility in numerals and English, respectively. These findings offer actionable insights for optimizing interfaces with different character types and highlight the need to explore further how display factors contribute to visual experience in complex scripts.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 103019"},"PeriodicalIF":3.7,"publicationDate":"2025-03-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143591642","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
vAQA-SS: Vision-based action quality assessment for style-based skiing
IF 3.7 2区 工程技术 Q1 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-03-04 DOI: 10.1016/j.displa.2025.103020
Yijia Wen , Xiaoyan Luo , Lei Zheng , Liangnan Qi , Xiaofeng Shi
Vision-based Action Quality Assessment (AQA) aims to evaluate action quality in video data, aligning with the subjective scores of human experts. Due to the unique challenges posed by different sports, it is difficult to design a uniform AQA system applicable to all sports. Consequently, many current sports AQA methods focus on specific disciplines such as diving and gymnastics. In contrast, skiing AQA, characterized by high-dynamic actions and complex outdoor scenes, faces additional challenges. Therefore, we constructed a specific dataset for style-based skiing, which focuses athlete’s movement style and execution, encompassing diverse skiing events with detailed annotations on action classes and athletes’ final scores, named Skiing-6. Leveraging this dataset, we designed two vision-based skiing action quality assessment (vAQA-SS) models. One model directly generates an absolute AQA score by measuring the quality of an athlete’s actions in the input video without any external reference, termed orAQA, which assesses athlete performance based on low-level spatiotemporal features of the video data alongside high-level pose features. The other model calculates a relative AQA score, deriving the performance score of an athlete’s actions from the source input video with a reference video, termed wrAQA. Finally, we conducted extensive experiments on Skiing-6 and SkiTB to demonstrate the effectiveness of our vAQA-SS models. The results demonstrate that our approach achieves significant improvements in both absolute evaluation (orAQA) and relative evaluation (wrAQA), surpassing other similar sports AQA methods.
{"title":"vAQA-SS: Vision-based action quality assessment for style-based skiing","authors":"Yijia Wen ,&nbsp;Xiaoyan Luo ,&nbsp;Lei Zheng ,&nbsp;Liangnan Qi ,&nbsp;Xiaofeng Shi","doi":"10.1016/j.displa.2025.103020","DOIUrl":"10.1016/j.displa.2025.103020","url":null,"abstract":"<div><div>Vision-based Action Quality Assessment (AQA) aims to evaluate action quality in video data, aligning with the subjective scores of human experts. Due to the unique challenges posed by different sports, it is difficult to design a uniform AQA system applicable to all sports. Consequently, many current sports AQA methods focus on specific disciplines such as diving and gymnastics. In contrast, skiing AQA, characterized by high-dynamic actions and complex outdoor scenes, faces additional challenges. Therefore, we constructed a specific dataset for style-based skiing, which focuses athlete’s movement style and execution, encompassing diverse skiing events with detailed annotations on action classes and athletes’ final scores, named Skiing-6. Leveraging this dataset, we designed two vision-based skiing action quality assessment (vAQA-SS) models. One model directly generates an absolute AQA score by measuring the quality of an athlete’s actions in the input video without any external reference, termed <em>or</em>AQA, which assesses athlete performance based on low-level spatiotemporal features of the video data alongside high-level pose features. The other model calculates a relative AQA score, deriving the performance score of an athlete’s actions from the source input video with a reference video, termed <em>wr</em>AQA. Finally, we conducted extensive experiments on Skiing-6 and SkiTB to demonstrate the effectiveness of our vAQA-SS models. The results demonstrate that our approach achieves significant improvements in both absolute evaluation (<em>or</em>AQA) and relative evaluation (<em>wr</em>AQA), surpassing other similar sports AQA methods.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 103020"},"PeriodicalIF":3.7,"publicationDate":"2025-03-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143642249","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adversarially Regularized Tri-Transformer Fusion for continual multimodal egocentric activity recognition
IF 3.7 2区 工程技术 Q1 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-02-27 DOI: 10.1016/j.displa.2025.102992
Shuchang Zhou , Hanxin Wang , Qingbo Wu , Fanman Meng , Linfeng Xu , Wei Zhang , Hongliang Li
Continual egocentric activity recognition aims to understand first-person activity from the multimodal data captured from wearable devices in streaming environments. Existing continual learning (CL) methods hardly acquire discriminative multimodal representations of activity classes from different isolated stages. To address this issue, this paper proposes an Adversarially Regularized Tri-Transformer Fusion (ARTF) model composed of three frozen transformer backbones with dynamic expansion architecture, which enables flexible and progressive multimodal representation fusion in the CL setting. To mitigate the confusion across different stages, we adopt an adversary-based confusion feature generation strategy to augment unknown classes, explicitly simulating out-stage features that closely resemble those within the stage. Then, the discriminative multimodal fusion representations could be learned by joint training on the current and augmented data at different stages. Experiments show that our model significantly outperforms state-of-the-art CL methods for multimodal continual egocentric activity recognition.
{"title":"Adversarially Regularized Tri-Transformer Fusion for continual multimodal egocentric activity recognition","authors":"Shuchang Zhou ,&nbsp;Hanxin Wang ,&nbsp;Qingbo Wu ,&nbsp;Fanman Meng ,&nbsp;Linfeng Xu ,&nbsp;Wei Zhang ,&nbsp;Hongliang Li","doi":"10.1016/j.displa.2025.102992","DOIUrl":"10.1016/j.displa.2025.102992","url":null,"abstract":"<div><div>Continual egocentric activity recognition aims to understand first-person activity from the multimodal data captured from wearable devices in streaming environments. Existing continual learning (CL) methods hardly acquire discriminative multimodal representations of activity classes from different isolated stages. To address this issue, this paper proposes an Adversarially Regularized Tri-Transformer Fusion (ARTF) model composed of three frozen transformer backbones with dynamic expansion architecture, which enables flexible and progressive multimodal representation fusion in the CL setting. To mitigate the confusion across different stages, we adopt an adversary-based confusion feature generation strategy to augment unknown classes, explicitly simulating out-stage features that closely resemble those within the stage. Then, the discriminative multimodal fusion representations could be learned by joint training on the current and augmented data at different stages. Experiments show that our model significantly outperforms state-of-the-art CL methods for multimodal continual egocentric activity recognition.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 102992"},"PeriodicalIF":3.7,"publicationDate":"2025-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143529734","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A general and flexible point cloud simplification method based on feature fusion 基于特征融合的通用灵活点云简化方法
IF 3.7 2区 工程技术 Q1 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2025-02-26 DOI: 10.1016/j.displa.2025.103007
Jiale Chao, Jialin Lei, Xionghui Zhou, Le Xie
Large-scale, high-density point cloud data often pose challenges for direct application in various downstream tasks. To address this issue, this paper introduces a flexible point cloud simplification method based on feature fusion. After conducting a comprehensive analysis of the input point cloud, the method fuses the density feature that reflects point cloud uniformity with local geometric features that capture shape details. Based on the simplification objectives and fused feature values, the method optimizes the point distribution from a global perspective. Subsequently, by removing distance factors, purely local geometric features are incorporated into the farthest point sampling process and a feature-weighted voxel farthest point sampling algorithm is proposed to prioritize the preservation of local feature points. With a refined mechanism for adjusting point numbers, the method finally achieves fast and reasonable simplification of massive point clouds. Furthermore, extensive experiments have been designed to explore the impact of the features involved and their sensitivity to simplification results, offering detailed recommendations for parameter configuration. This method supports flexible transitions between global uniformity and heavy local feature preservation. Comparative results with previous studies demonstrate its excellent balance, exhibiting strong competitiveness in both output point cloud quality and computational efficiency. The core source code is publicly available at: https://github.com/chaojiale/PointCloudSimplification.
{"title":"A general and flexible point cloud simplification method based on feature fusion","authors":"Jiale Chao,&nbsp;Jialin Lei,&nbsp;Xionghui Zhou,&nbsp;Le Xie","doi":"10.1016/j.displa.2025.103007","DOIUrl":"10.1016/j.displa.2025.103007","url":null,"abstract":"<div><div>Large-scale, high-density point cloud data often pose challenges for direct application in various downstream tasks. To address this issue, this paper introduces a flexible point cloud simplification method based on feature fusion. After conducting a comprehensive analysis of the input point cloud, the method fuses the density feature that reflects point cloud uniformity with local geometric features that capture shape details. Based on the simplification objectives and fused feature values, the method optimizes the point distribution from a global perspective. Subsequently, by removing distance factors, purely local geometric features are incorporated into the farthest point sampling process and a feature-weighted voxel farthest point sampling algorithm is proposed to prioritize the preservation of local feature points. With a refined mechanism for adjusting point numbers, the method finally achieves fast and reasonable simplification of massive point clouds. Furthermore, extensive experiments have been designed to explore the impact of the features involved and their sensitivity to simplification results, offering detailed recommendations for parameter configuration. This method supports flexible transitions between global uniformity and heavy local feature preservation. Comparative results with previous studies demonstrate its excellent balance, exhibiting strong competitiveness in both output point cloud quality and computational efficiency. The core source code is publicly available at: <span><span>https://github.com/chaojiale/PointCloudSimplification</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 103007"},"PeriodicalIF":3.7,"publicationDate":"2025-02-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143529733","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Displays
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1