Pub Date : 2024-08-17DOI: 10.1016/j.ijhcs.2024.103357
This study introduces an empirical approach for assessing human scent-related experiences within the field of Human-Computer Interaction (HCI). We labeled 43 fragrances based on grounded collective experience, incorporating semantic and impression-based data. Furthermore, we collected comprehensive psychophysiological data, including electroencephalogram (EEG), electrobulbogram (EBG), electrocardiogram (ECG), and facial dynamics captured by a camera, from participants who experienced the scents. By computing scent-wise similarity and correlating both grounded and psychophysiological scent spaces, we identified associations between them, demonstrating the potential of this approach to enhance our understanding of scent-related experiences. Additionally, we propose an iterative evaluation framework to refine the design of smell-based interactions and we conduct a real-life study to validate this framework.
{"title":"Representing scents: An evaluation framework of scent-related experiences through associations between grounded and psychophysiological data","authors":"","doi":"10.1016/j.ijhcs.2024.103357","DOIUrl":"10.1016/j.ijhcs.2024.103357","url":null,"abstract":"<div><p>This study introduces an empirical approach for assessing human scent-related experiences within the field of Human-Computer Interaction (HCI). We labeled 43 fragrances based on grounded collective experience, incorporating semantic and impression-based data. Furthermore, we collected comprehensive psychophysiological data, including electroencephalogram (EEG), electrobulbogram (EBG), electrocardiogram (ECG), and facial dynamics captured by a camera, from participants who experienced the scents. By computing scent-wise similarity and correlating both grounded and psychophysiological scent spaces, we identified associations between them, demonstrating the potential of this approach to enhance our understanding of scent-related experiences. Additionally, we propose an iterative evaluation framework to refine the design of smell-based interactions and we conduct a real-life study to validate this framework.</p></div>","PeriodicalId":54955,"journal":{"name":"International Journal of Human-Computer Studies","volume":null,"pages":null},"PeriodicalIF":5.3,"publicationDate":"2024-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142088291","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-16DOI: 10.1016/j.ijhcs.2024.103354
Fluency with computer applications has assumed a crucial role in work-related and other day-to-day activities. While prior experience is known to predict performance in tasks involving computers, the effects of more stable factors like cognitive abilities remain unclear. Here, we report findings from a controlled study () covering a wide spectrum of commonplace applications, from spreadsheets to video conferencing. Our main result is that cognitive abilities exert a significant, independent, and broad-based effect on computer users’ performance. In particular, users with high working memory, executive control, and perceptual reasoning ability complete tasks more quickly and with greater success while experiencing lower mental load. Remarkably, these effects are similar to or even larger in magnitude than the effects of prior experience in using computers and in completing tasks similar to those encountered in our study. However, the effects are varying and application-specific. We discuss the role that user interface design bears on decreasing ability-related differences, alongside benefits this could yield for functioning in society.
{"title":"Cognitive abilities predict performance in everyday computer tasks","authors":"","doi":"10.1016/j.ijhcs.2024.103354","DOIUrl":"10.1016/j.ijhcs.2024.103354","url":null,"abstract":"<div><p>Fluency with computer applications has assumed a crucial role in work-related and other day-to-day activities. While prior experience is known to predict performance in tasks involving computers, the effects of more stable factors like cognitive abilities remain unclear. Here, we report findings from a controlled study (<span><math><mrow><mi>N</mi><mo>=</mo><mn>88</mn></mrow></math></span>) covering a wide spectrum of commonplace applications, from spreadsheets to video conferencing. Our main result is that cognitive abilities exert a significant, independent, and broad-based effect on computer users’ performance. In particular, users with high working memory, executive control, and perceptual reasoning ability complete tasks more quickly and with greater success while experiencing lower mental load. Remarkably, these effects are similar to or even larger in magnitude than the effects of prior experience in using computers and in completing tasks similar to those encountered in our study. However, the effects are varying and application-specific. We discuss the role that user interface design bears on decreasing ability-related differences, alongside benefits this could yield for functioning in society.</p></div>","PeriodicalId":54955,"journal":{"name":"International Journal of Human-Computer Studies","volume":null,"pages":null},"PeriodicalIF":5.3,"publicationDate":"2024-08-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S107158192400137X/pdfft?md5=a902433d6ce6aad8ad7b4833a2deb786&pid=1-s2.0-S107158192400137X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142048239","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-14DOI: 10.1016/j.ijhcs.2024.103356
Background
A main challenge in many types of physical rehabilitation is patient adherence to recommended exercises. Vestibular rehabilitation is the most effective treatment for the symptoms of dizziness, vertigo, imbalance, and nausea caused by vestibular disorders, but adherence levels are particularly low as the rehabilitation program calls for many short exercise sets during the day, which can worsen symptoms and impair balance in the short term. Technological tools have the potential to increase adherence, but to date, there has been no comprehensive analysis, in the context of vestibular rehabilitation, of the specific needs from technology, of its limitations, and of concerns regarding its use.
Objective
The aim of the study is to identify the main features required from technology for vestibular rehabilitation, as perceived by patients with vestibular disorders and by vestibular physical therapists, using a socially assistive robot as a test case. We seek here to provide practical information for the development of future vestibular rehabilitation technologies which are based on human-computer interaction (HCI) and human-robot interaction (HRI).
Methods
We conducted a qualitative study with six focus groups (N = 39). Three groups of patients with vestibular disorders (N = 18) and three groups of physical therapists (N = 21) participated in this study. The participants answered structured questions on technologies for vestibular rehabilitation, watched a presentation of two videos of a socially assistive robot (SAR), and completed an online survey. Thematic analysis with a mixed deductive and inductive approach was used to analyze the data.
Results
Participants preferred phone applications or virtual/augmented reality platforms over an embodied robotic platform. They wanted technology to be adaptive to the different stages of rehabilitation, gamified, easy to use, safe, reliable, portable, and accessible remotely by the therapist. They reported that the technology should provide feedback on the quality and quantity of exercise performance and monitor these factors while considering the tolerability of the ensuing disruptive symptoms. Participants expected that using technology as part of the rehabilitation process would shorten exercise sessions and improve clinical outcomes compared to standard care. SARs for vestibular rehabilitation were perceived as useful mostly for children and patients with chronic vestibular disorders, and their potential use for rehabilitation raised concerns regarding safety, ethics, and technical complexity.
Conclusions
Although SARs can potentially be used to increase exercise adherence, a phone application appears to be a more suitable medium for this purpose, raising fewer notable concerns from users. We provide a summary of perceived advantages and disadvantages of te
{"title":"Do we really need this robot? Technology requirements for vestibular rehabilitation: Input from patients and clinicians","authors":"","doi":"10.1016/j.ijhcs.2024.103356","DOIUrl":"10.1016/j.ijhcs.2024.103356","url":null,"abstract":"<div><h3>Background</h3><p>A main challenge in many types of physical rehabilitation is patient adherence to recommended exercises. Vestibular rehabilitation is the most effective treatment for the symptoms of dizziness, vertigo, imbalance, and nausea caused by vestibular disorders, but adherence levels are particularly low as the rehabilitation program calls for many short exercise sets during the day, which can worsen symptoms and impair balance in the short term. Technological tools have the potential to increase adherence, but to date, there has been no comprehensive analysis, in the context of vestibular rehabilitation, of the specific needs from technology, of its limitations, and of concerns regarding its use.</p></div><div><h3>Objective</h3><p>The aim of the study is to identify the main features required from technology for vestibular rehabilitation, as perceived by patients with vestibular disorders and by vestibular physical therapists, using a socially assistive robot as a test case. We seek here to provide practical information for the development of future vestibular rehabilitation technologies which are based on human-computer interaction (HCI) and human-robot interaction (HRI).</p></div><div><h3>Methods</h3><p>We conducted a qualitative study with six focus groups (<em>N</em> = 39). Three groups of patients with vestibular disorders (<em>N</em> = 18) and three groups of physical therapists (<em>N</em> = 21) participated in this study. The participants answered structured questions on technologies for vestibular rehabilitation, watched a presentation of two videos of a socially assistive robot (SAR), and completed an online survey. Thematic analysis with a mixed deductive and inductive approach was used to analyze the data.</p></div><div><h3>Results</h3><p>Participants preferred phone applications or virtual/augmented reality platforms over an embodied robotic platform. They wanted technology to be adaptive to the different stages of rehabilitation, gamified, easy to use, safe, reliable, portable, and accessible remotely by the therapist. They reported that the technology should provide feedback on the quality and quantity of exercise performance and monitor these factors while considering the tolerability of the ensuing disruptive symptoms. Participants expected that using technology as part of the rehabilitation process would shorten exercise sessions and improve clinical outcomes compared to standard care. SARs for vestibular rehabilitation were perceived as useful mostly for children and patients with chronic vestibular disorders, and their potential use for rehabilitation raised concerns regarding safety, ethics, and technical complexity.</p></div><div><h3>Conclusions</h3><p>Although SARs can potentially be used to increase exercise adherence, a phone application appears to be a more suitable medium for this purpose, raising fewer notable concerns from users. We provide a summary of perceived advantages and disadvantages of te","PeriodicalId":54955,"journal":{"name":"International Journal of Human-Computer Studies","volume":null,"pages":null},"PeriodicalIF":5.3,"publicationDate":"2024-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S1071581924001393/pdfft?md5=d6491755bf4e3baa08ca08cd42cb3db8&pid=1-s2.0-S1071581924001393-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142048237","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-13DOI: 10.1016/j.ijhcs.2024.103355
With the expansive growth of AI’s capabilities in recent years, researchers have been tasked with developing and improving human-centered AI collaborations, necessitating the creation of human–AI teams (HATs). However, the differences in communication styles between humans and AI often prevent human teammates from fully understanding the intent and needs of AI teammates. One core difference is that humans naturally leverage a positive emotional tone during communication to convey their confidence or lack thereof to convey doubt in their ability to complete a task. Yet, this communication strategy must be explicitly designed in order for an AI teammate to be human-centered. In this mixed-methods study, 45 participants completed a study examining how human teammates interpret the behaviors of their AI teammates when they express different positive emotions via specific words/phrases. Quantitative results show that, based on corresponding behaviors, AI teammates were able to use displays of emotion to increase trust in AI teammates and the positive mood of the human teammate. Additionally, our qualitative findings indicate that participants preferred their AI teammates to increase the intensity of their displayed emotions to help reduce the perceived risk of their AI teammate’s behavior. When taken in sum, these findings describe the relevance of AI teammates expressing intensities of emotion while performing various behavioral decisions as a continued means to provide social support to the wider team and better task performance.
{"title":"What you say vs what you do: Utilizing positive emotional expressions to relay AI teammate intent within human–AI teams","authors":"","doi":"10.1016/j.ijhcs.2024.103355","DOIUrl":"10.1016/j.ijhcs.2024.103355","url":null,"abstract":"<div><p>With the expansive growth of AI’s capabilities in recent years, researchers have been tasked with developing and improving human-centered AI collaborations, necessitating the creation of human–AI teams (HATs). However, the differences in communication styles between humans and AI often prevent human teammates from fully understanding the intent and needs of AI teammates. One core difference is that humans naturally leverage a positive emotional tone during communication to convey their confidence or lack thereof to convey doubt in their ability to complete a task. Yet, this communication strategy must be explicitly designed in order for an AI teammate to be human-centered. In this mixed-methods study, 45 participants completed a study examining how human teammates interpret the behaviors of their AI teammates when they express different positive emotions via specific words/phrases. Quantitative results show that, based on corresponding behaviors, AI teammates were able to use displays of emotion to increase trust in AI teammates and the positive mood of the human teammate. Additionally, our qualitative findings indicate that participants preferred their AI teammates to increase the intensity of their displayed emotions to help reduce the perceived risk of their AI teammate’s behavior. When taken in sum, these findings describe the relevance of AI teammates expressing intensities of emotion while performing various behavioral decisions as a continued means to provide social support to the wider team and better task performance.</p></div>","PeriodicalId":54955,"journal":{"name":"International Journal of Human-Computer Studies","volume":null,"pages":null},"PeriodicalIF":5.3,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142011569","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-13DOI: 10.1016/j.ijhcs.2024.103347
Micro-electronics tools, coupled with card-based tools, are employed for prototyping smart devices with non-experts. Lately, researchers have started investigating what tools can actively engage people with intellectual disabilities (ID) in their prototyping. This paper posits itself in this line of work. It presents a toolkit for ID people to rapidly prototype together their own ideas of smart things, for their own shared environment. It analyses and discusses engaging or disengaging features of the toolkit in light of the results of two workshops with eight ID participants. Lessons of broad interest for the design of similar toolkits are drawn from the literature and study findings.
微电子工具和基于卡片的工具被用于为非专业人员制作智能设备原型。最近,研究人员开始研究什么样的工具能让智障人士积极参与原型设计。本文就属于这一研究领域。它为智障人士提供了一个工具包,让他们在自己的共享环境中,快速将自己对智能事物的想法制作成原型。本文根据与八位 ID 参与者进行的两次研讨会的结果,分析并讨论了该工具包的吸引人或不吸引人之处。从文献和研究结果中汲取了对设计类似工具包具有广泛意义的经验教训。
{"title":"A rapid-prototyping toolkit for people with intellectual disabilities","authors":"","doi":"10.1016/j.ijhcs.2024.103347","DOIUrl":"10.1016/j.ijhcs.2024.103347","url":null,"abstract":"<div><p>Micro-electronics tools, coupled with card-based tools, are employed for prototyping smart devices with non-experts. Lately, researchers have started investigating what tools can actively engage people with intellectual disabilities (ID) in their prototyping. This paper posits itself in this line of work. It presents a toolkit for ID people to rapidly prototype together their own ideas of smart things, for their own shared environment. It analyses and discusses engaging or disengaging features of the toolkit in light of the results of two workshops with eight ID participants. Lessons of broad interest for the design of similar toolkits are drawn from the literature and study findings.</p></div>","PeriodicalId":54955,"journal":{"name":"International Journal of Human-Computer Studies","volume":null,"pages":null},"PeriodicalIF":5.3,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142076253","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-10DOI: 10.1016/j.ijhcs.2024.103345
Fall detection cameras at home can detect emergencies of older adults and send timely life-saving alerts. However, the equilibrium between privacy protection and life safety remains a controversial issue when using cameras. In this study, we assessed the attitudes of older adults towards the privacy issue of cameras using surveys (N=389) and interviews (N=20). Furthermore, we conducted a co-design workshop (N=6) in which older adults and designers collaborated to develop a prototype of cameras. We found that for older adults, the disclosure of privacy not only involves a leakage of personal information, but also influences their dignity and control, which has rarely been expressed directly in the past. Our results expand the conceptualisation of privacy and provide novel design implications for smart product development on privacy for older adults.
{"title":"Beyond digital privacy: Uncovering deeper attitudes toward privacy in cameras among older adults","authors":"","doi":"10.1016/j.ijhcs.2024.103345","DOIUrl":"10.1016/j.ijhcs.2024.103345","url":null,"abstract":"<div><p>Fall detection cameras at home can detect emergencies of older adults and send timely life-saving alerts. However, the equilibrium between privacy protection and life safety remains a controversial issue when using cameras. In this study, we assessed the attitudes of older adults towards the privacy issue of cameras using surveys (N=389) and interviews (N=20). Furthermore, we conducted a co-design workshop (N=6) in which older adults and designers collaborated to develop a prototype of cameras. We found that for older adults, the disclosure of privacy not only involves a leakage of personal information, but also influences their dignity and control, which has rarely been expressed directly in the past. Our results expand the conceptualisation of privacy and provide novel design implications for smart product development on privacy for older adults.</p></div>","PeriodicalId":54955,"journal":{"name":"International Journal of Human-Computer Studies","volume":null,"pages":null},"PeriodicalIF":5.3,"publicationDate":"2024-08-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142076254","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-07DOI: 10.1016/j.ijhcs.2024.103342
Virtual Reality (VR) enables the low-cost production of realistic prototypes of buildings at early stages of architectural projects. Such prototypes may be used to gather the experiences of future users and iterate early on in the design. However, it is essential to evaluate whether what is experienced within such VR prototypes corresponds to what will be experienced in reality. Here, we use an innovative method to compare the experiences of patients in a real building and in a virtual environment that plays the role of a prototype that could have been created by architects during the design phase. We first designed and implemented a VR environment replicating an existing ambulatory pathway. Then, we used micro-phenomenological interviews to collect the experiences of real patients in the VR environment (n=8), along with VR traces and first-person point of view videos, and in the real ambulatory pathway (n=8). We modeled and normalized the experiences, and compared them systematically. Results suggest that patients live comparable experiences along various experiential dimensions such as thought, emotion, sensation, social and sensory perceptions, and that VR prototypes may be adequate to assess issues with architectural design. This work opens unique perspectives towards involving patients in User-Centered Design in architecture, though challenges lie ahead in how to design VR prototypes from early blueprints of architects.
{"title":"Towards the use of virtual reality prototypes in architecture to collect user experiences: An assessment of the comparability of patient experiences in a virtual and a real ambulatory pathway","authors":"","doi":"10.1016/j.ijhcs.2024.103342","DOIUrl":"10.1016/j.ijhcs.2024.103342","url":null,"abstract":"<div><p>Virtual Reality (VR) enables the low-cost production of realistic prototypes of buildings at early stages of architectural projects. Such prototypes may be used to gather the experiences of future users and iterate early on in the design. However, it is essential to evaluate whether what is experienced within such VR prototypes corresponds to what will be experienced in reality. Here, we use an innovative method to compare the experiences of patients in a real building and in a virtual environment that plays the role of a prototype that could have been created by architects during the design phase. We first designed and implemented a VR environment replicating an existing ambulatory pathway. Then, we used micro-phenomenological interviews to collect the experiences of real patients in the VR environment (n=8), along with VR traces and first-person point of view videos, and in the real ambulatory pathway (n=8). We modeled and normalized the experiences, and compared them systematically. Results suggest that patients live comparable experiences along various experiential dimensions such as thought, emotion, sensation, social and sensory perceptions, and that VR prototypes may be adequate to assess issues with architectural design. This work opens unique perspectives towards involving patients in User-Centered Design in architecture, though challenges lie ahead in how to design VR prototypes from early blueprints of architects.</p></div>","PeriodicalId":54955,"journal":{"name":"International Journal of Human-Computer Studies","volume":null,"pages":null},"PeriodicalIF":5.3,"publicationDate":"2024-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S1071581924001253/pdfft?md5=e68b25d62c0e0391983e65bcbf8c4366&pid=1-s2.0-S1071581924001253-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142088289","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-30DOI: 10.1016/j.ijhcs.2024.103341
This paper investigates the relationship between design research and humancomputer interaction (HCI) in the context of climate change communication and engagement. We discuss current practices in climate change communication and the decrease in concern and engagement caused by “crisis fatigue”. Through Research through Design (RtD), we set out to investigate data humanism and how users react to climate change data, testing approaches to improve engagement. With this purpose, we designed and evaluated Finding Arcadia, an interactive data story that uses data humanism to shift the dialogue from crisis-focused to action-focused. One study with the original IMF visualisations (N = 17) and two studies in public spaces (N = 12 and N = 64) point to the contextualization of the data and presenting actionable solutions helping in engaging users with climate change issues; help in creating solution-focused narratives and interpreting and relating with climate data. From these results, we derive insights for designing empowering interactive data visualizations for resilient climate change engagement.
本文以气候变化传播和参与为背景,探讨了设计研究与人机交互(HCI)之间的关系。我们讨论了气候变化传播的当前实践,以及 "危机疲劳 "导致的关注度和参与度下降。通过设计研究(RtD),我们着手调查数据人文主义以及用户如何对气候变化数据做出反应,测试提高参与度的方法。为此,我们设计并评估了 "寻找阿卡迪亚"(Finding Arcadia),这是一个互动数据故事,利用数据人文主义将对话从关注危机转向关注行动。一项关于原始 IMF 可视化的研究(N = 17)和两项关于公共空间的研究(N = 12 和 N = 64)表明,数据的背景化和提出可操作的解决方案有助于吸引用户参与气候变化问题;有助于创建以解决方案为重点的叙事以及解释气候数据并与之建立联系。从这些结果中,我们得出了设计赋权互动数据可视化的见解,以促进有弹性的气候变化参与。
{"title":"Connecting audiences with climate change: Towards humanised and action-focused data interactions","authors":"","doi":"10.1016/j.ijhcs.2024.103341","DOIUrl":"10.1016/j.ijhcs.2024.103341","url":null,"abstract":"<div><p>This paper investigates the relationship between design research and humancomputer interaction (HCI) in the context of climate change communication and engagement. We discuss current practices in climate change communication and the decrease in concern and engagement caused by “crisis fatigue”. Through Research through Design (RtD), we set out to investigate data humanism and how users react to climate change data, testing approaches to improve engagement. With this purpose, we designed and evaluated <em>Finding Arcadia</em>, an interactive data story that uses data humanism to shift the dialogue from crisis-focused to action-focused. One study with the original IMF visualisations (<em>N</em> = 17) and two studies in public spaces (<em>N</em> = 12 and <em>N</em> = 64) point to the contextualization of the data and presenting actionable solutions helping in engaging users with climate change issues; help in creating solution-focused narratives and interpreting and relating with climate data. From these results, we derive insights for designing empowering interactive data visualizations for resilient climate change engagement.</p></div>","PeriodicalId":54955,"journal":{"name":"International Journal of Human-Computer Studies","volume":null,"pages":null},"PeriodicalIF":5.3,"publicationDate":"2024-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S1071581924001241/pdfft?md5=a32ba3c151f3843aa67936d8aa9a2206&pid=1-s2.0-S1071581924001241-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141963845","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-26DOI: 10.1016/j.ijhcs.2024.103344
Considering that a significant portion of the current pedestrian population has limited exposure to automated vehicles (AVs), it is crucial to have a reliable instrument for assessing pedestrians’ initial trust in AVs. Using a survey of 436 pedestrians, this study developed and validated a PITQA (Pedestrians’ Initial Trust Questionnaire for AVs) scale using partial least squares structural equation modeling (PLS-SEM). The proposed scale will be valuable in monitoring the progression of trust over time and considering trust-related factors during the design process. The results revealed that seven key constructs significantly contribute to predicting initial trust between pedestrians and AVs. These constructs include propensity to trust, perceived statistical reliability, dependability and competence, perceived predictability, familiarity, authority/subversion, care/harm, and sanctity/degradation. These shed light on how the trust propensity of individuals, different trust/trustworthiness attributes might constitute different aspects of initial trust in the pedestrian-AV context. The developed scale can be a potentially useful tool for future research endeavors concerning trust calibration and the design of AVs specifically tailored for vulnerable road users.
{"title":"Development of a measurement instrument for pedestrians’ initial trust in automated vehicles","authors":"","doi":"10.1016/j.ijhcs.2024.103344","DOIUrl":"10.1016/j.ijhcs.2024.103344","url":null,"abstract":"<div><p>Considering that a significant portion of the current pedestrian population has limited exposure to automated vehicles (AVs), it is crucial to have a reliable instrument for assessing pedestrians’ initial trust in AVs. Using a survey of 436 pedestrians, this study developed and validated a PITQA (Pedestrians’ Initial Trust Questionnaire for AVs) scale using partial least squares structural equation modeling (PLS-SEM). The proposed scale will be valuable in monitoring the progression of trust over time and considering trust-related factors during the design process. The results revealed that seven key constructs significantly contribute to predicting initial trust between pedestrians and AVs. These constructs include <em>propensity to trust, perceived statistical reliability, dependability and competence, perceived predictability, familiarity, authority/subversion, care/harm</em>, and <em>sanctity/degradation</em>. These shed light on how the trust propensity of individuals, different trust/trustworthiness attributes might constitute different aspects of initial trust in the pedestrian-AV context. The developed scale can be a potentially useful tool for future research endeavors concerning trust calibration and the design of AVs specifically tailored for vulnerable road users.</p></div>","PeriodicalId":54955,"journal":{"name":"International Journal of Human-Computer Studies","volume":null,"pages":null},"PeriodicalIF":5.3,"publicationDate":"2024-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S1071581924001277/pdfft?md5=b2bb29c21ca7b7eafefd0b998c22c85c&pid=1-s2.0-S1071581924001277-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141851092","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-23DOI: 10.1016/j.ijhcs.2024.103340
The integration of emotion recognition capabilities within musical instruments can spur the emergence of novel art formats and services for musicians. This paper proposes the concept of emotionally-aware smart musical instruments, a class of musical devices embedding an artificial intelligence agent able to recognize the emotion contained in the musical signal. This spurs the emergence of novel services for musicians. Two prototypes of emotionally-aware smart piano and smart electric guitar were created, which embedded a recognition method for happiness, sadness, relaxation, aggressiveness and combination thereof. A user study, conducted with eleven pianists and eleven electric guitarists, revealed the strengths and limitations of the developed technology. On average musicians appreciated the proposed concept, who found its value in various musical activities. Most of participants tended to justify the system with respect to erroneous or partially erroneous classifications of the emotions they expressed, reporting to understand the reasons why a given output was produced. Some participants even seemed to trust more the system than their own judgments. Conversely, other participants requested to improve the accuracy, reliability and explainability of the system in order to achieve a higher degree of partnership with it. Our results suggest that, while desirable, perfect prediction of the intended emotion is not an absolute requirement for music emotion recognition to be useful in the construction of smart musical instruments.
{"title":"Musician-AI partnership mediated by emotionally-aware smart musical instruments","authors":"","doi":"10.1016/j.ijhcs.2024.103340","DOIUrl":"10.1016/j.ijhcs.2024.103340","url":null,"abstract":"<div><p>The integration of emotion recognition capabilities within musical instruments can spur the emergence of novel art formats and services for musicians. This paper proposes the concept of emotionally-aware smart musical instruments, a class of musical devices embedding an artificial intelligence agent able to recognize the emotion contained in the musical signal. This spurs the emergence of novel services for musicians. Two prototypes of emotionally-aware smart piano and smart electric guitar were created, which embedded a recognition method for happiness, sadness, relaxation, aggressiveness and combination thereof. A user study, conducted with eleven pianists and eleven electric guitarists, revealed the strengths and limitations of the developed technology. On average musicians appreciated the proposed concept, who found its value in various musical activities. Most of participants tended to justify the system with respect to erroneous or partially erroneous classifications of the emotions they expressed, reporting to understand the reasons why a given output was produced. Some participants even seemed to trust more the system than their own judgments. Conversely, other participants requested to improve the accuracy, reliability and explainability of the system in order to achieve a higher degree of partnership with it. Our results suggest that, while desirable, perfect prediction of the intended emotion is not an absolute requirement for music emotion recognition to be useful in the construction of smart musical instruments.</p></div>","PeriodicalId":54955,"journal":{"name":"International Journal of Human-Computer Studies","volume":null,"pages":null},"PeriodicalIF":5.3,"publicationDate":"2024-07-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S107158192400123X/pdfft?md5=9a551949f1594cc78460b20e32ef1a41&pid=1-s2.0-S107158192400123X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141850114","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}