首页 > 最新文献

Adv. Hum. Comput. Interact.最新文献

英文 中文
Vision-Based Interfaces for Character-Based Text Entry: Comparison of Errors and Error Correction Properties of Eye Typing and Head Typing 基于视觉的字符文本输入界面:眼动打字和头动打字的错误和纠错特性比较
Pub Date : 2023-11-22 DOI: 10.1155/2023/8855764
Y. Gizatdinova, O. Špakov, O. Tuisku, Matthew Turk, Veikko Surakka
We examined two vision-based interfaces (VBIs) for performance and user experience during character-based text entry using an on-screen virtual keyboard. Head-based VBI uses head motion to steer the computer pointer and mouth-opening gestures to select the keyboard keys. Gaze-based VBI utilizes gaze for pointing at the keys and an adjustable dwell for key selection. The results showed that after three sessions (45 min of typing in total), able-bodied novice participants (N = 34) typed significantly slower yet yielded significantly more accurate text with head-based VBI with gaze-based VBIs. The analysis of errors and corrective actions relative to the spatial layout of the keyboard revealed a difference in the error correction behavior of the participants when typing using both interfaces. We estimated the error correction cost for both interfaces and suggested implications for the future use and improvement of VBIs for hands-free text entry.
我们研究了两种基于视觉的界面(VBI),以了解使用屏幕虚拟键盘输入基于字符的文本时的性能和用户体验。基于头部的 VBI 使用头部运动来引导计算机指针,并通过张嘴手势来选择键盘按键。基于注视的 VBI 利用注视来指向按键,并利用可调节的停留时间来选择按键。结果表明,经过三次训练(共输入 45 分钟)后,身体健全的新手参与者(34 人)在使用基于头部的 VBI 和基于注视的 VBI 时,输入速度明显较慢,但输入文本的准确性却明显更高。通过分析与键盘空间布局相关的错误和纠正措施,我们发现参与者在使用这两种界面打字时的纠错行为存在差异。我们估算了两种界面的纠错成本,并对未来使用和改进免提文字输入的虚拟生物识别技术提出了建议。
{"title":"Vision-Based Interfaces for Character-Based Text Entry: Comparison of Errors and Error Correction Properties of Eye Typing and Head Typing","authors":"Y. Gizatdinova, O. Špakov, O. Tuisku, Matthew Turk, Veikko Surakka","doi":"10.1155/2023/8855764","DOIUrl":"https://doi.org/10.1155/2023/8855764","url":null,"abstract":"We examined two vision-based interfaces (VBIs) for performance and user experience during character-based text entry using an on-screen virtual keyboard. Head-based VBI uses head motion to steer the computer pointer and mouth-opening gestures to select the keyboard keys. Gaze-based VBI utilizes gaze for pointing at the keys and an adjustable dwell for key selection. The results showed that after three sessions (45 min of typing in total), able-bodied novice participants (N = 34) typed significantly slower yet yielded significantly more accurate text with head-based VBI with gaze-based VBIs. The analysis of errors and corrective actions relative to the spatial layout of the keyboard revealed a difference in the error correction behavior of the participants when typing using both interfaces. We estimated the error correction cost for both interfaces and suggested implications for the future use and improvement of VBIs for hands-free text entry.","PeriodicalId":192934,"journal":{"name":"Adv. Hum. Comput. Interact.","volume":"47 1","pages":"8855764:1-8855764:23"},"PeriodicalIF":0.0,"publicationDate":"2023-11-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139248982","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Learning Methods for Arabic Autoencoder Speech Recognition System for Electro-Larynx Device 用于电喉装置的阿拉伯语自动编码器语音识别系统的深度学习方法
Pub Date : 2023-02-28 DOI: 10.1155/2023/7398538
Z. J. M. Ameen, A. Kadhim
Recent advances in speech recognition have achieved remarkable performance comparable with human transcribers’ abilities. But this significant performance is not the same for all the spoken languages. The Arabic language is one of them. Arabic speech recognition is bounded to the lack of suitable datasets. Artificial intelligence algorithms have shown promising capabilities for Arabic speech recognition. Arabic is the official language of 22 countries, and it has been estimated that 400 million people speak the Arabic language worldwide. Speech disabilities have been one of the expanding problems in the last decades, even in kids. Some devices can be used to generate speech for those people. One of these devices is the Servox Digital Electro-Larynx (EL). In this research, we developed an autoencoder with a combination of long short-term memory (LSTM) and gated recurrent units (GRU) models to recognize recorded signals from Servox Digital EL Electro-Larynx. The proposed framework consisted of three steps: denoising, feature extraction, and Arabic speech recognition. The experimental results show 95.31% accuracy for Arabic speech recognition with the proposed model. In this research, we evaluated different combinations of LSTM and GRU for constructing the best autoencoder. A rigorous evaluation process indicates better performance with the use of GRU in both encoder and decoder structures. The proposed model achieved a 4.69% word error rate (WER). Experimental results confirm that the proposed model can be used for developing a real-time app to recognize common Arabic spoken words.
语音识别的最新进展已经取得了与人类转录能力相当的显著表现。但这种显著的表现并不适用于所有的口语。阿拉伯语就是其中之一。阿拉伯语语音识别受限于缺乏合适的数据集。人工智能算法在阿拉伯语语音识别方面表现出了很好的能力。阿拉伯语是22个国家的官方语言,据估计全世界有4亿人说阿拉伯语。在过去的几十年里,语言障碍一直是一个不断扩大的问题,甚至在儿童中也是如此。一些设备可以用来为这些人生成语音。其中一种装置是伺服数字电喉(EL)。在这项研究中,我们开发了一种结合长短期记忆(LSTM)和门控循环单元(GRU)模型的自编码器,以识别来自Servox Digital EL电喉的记录信号。该框架包括三个步骤:去噪、特征提取和阿拉伯语语音识别。实验结果表明,该模型对阿拉伯语语音识别的准确率为95.31%。在这项研究中,我们评估了LSTM和GRU的不同组合来构建最佳的自编码器。严格的评估过程表明,在编码器和解码器结构中使用GRU具有更好的性能。该模型的错误率为4.69%。实验结果表明,该模型可用于开发实时识别阿拉伯语常用口语单词的应用程序。
{"title":"Deep Learning Methods for Arabic Autoencoder Speech Recognition System for Electro-Larynx Device","authors":"Z. J. M. Ameen, A. Kadhim","doi":"10.1155/2023/7398538","DOIUrl":"https://doi.org/10.1155/2023/7398538","url":null,"abstract":"Recent advances in speech recognition have achieved remarkable performance comparable with human transcribers’ abilities. But this significant performance is not the same for all the spoken languages. The Arabic language is one of them. Arabic speech recognition is bounded to the lack of suitable datasets. Artificial intelligence algorithms have shown promising capabilities for Arabic speech recognition. Arabic is the official language of 22 countries, and it has been estimated that 400 million people speak the Arabic language worldwide. Speech disabilities have been one of the expanding problems in the last decades, even in kids. Some devices can be used to generate speech for those people. One of these devices is the Servox Digital Electro-Larynx (EL). In this research, we developed an autoencoder with a combination of long short-term memory (LSTM) and gated recurrent units (GRU) models to recognize recorded signals from Servox Digital EL Electro-Larynx. The proposed framework consisted of three steps: denoising, feature extraction, and Arabic speech recognition. The experimental results show 95.31% accuracy for Arabic speech recognition with the proposed model. In this research, we evaluated different combinations of LSTM and GRU for constructing the best autoencoder. A rigorous evaluation process indicates better performance with the use of GRU in both encoder and decoder structures. The proposed model achieved a 4.69% word error rate (WER). Experimental results confirm that the proposed model can be used for developing a real-time app to recognize common Arabic spoken words.","PeriodicalId":192934,"journal":{"name":"Adv. Hum. Comput. Interact.","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124778583","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
CHR vs. Human-Computer Interaction Design for Emerging Technologies: Two Case Studies 面向新兴技术的人机交互设计:两个案例研究
Pub Date : 2023-02-14 DOI: 10.1155/2023/8710638
Sharefa Murad, A. Qusef, Muhanna A. Muhanna
Recent years have seen a surge in interest in the multifaceted topic of human-computer interaction (HCI). Since the advent of the Fourth Industrial Revolution, the significance of human-computer interaction in the field of safety risk management has only grown. There has not been a lot of focus on developing human-computer interaction for identifying potential hazards in buildings. After conducting a comprehensive literature review, we developed a study framework for the use of human-computer interaction in the identification of construction-related hazards (CHR-HCI). Future studies will focus on the intersection of computer vision, VR, and ergonomics. In this research, we have built a theoretical foundation for past studies’ findings and connections and offered concrete recommendations for the improvement of HCI in danger identification in the future. Moreover, we analyzed two cases studies related to the domain of CHR-HCI in terms of wearable vibration-based systems and context aware navigation.
近年来,人们对人机交互(HCI)这一多面主题的兴趣激增。自第四次工业革命以来,人机交互在安全风险管理领域的重要性与日俱增。在开发人机交互以识别建筑物中的潜在危险方面,还没有得到很多关注。在进行了全面的文献综述之后,我们开发了一个研究框架,用于在建筑相关危害识别中使用人机交互(hr - hci)。未来的研究将集中于计算机视觉、虚拟现实和人体工程学的交叉。在本研究中,我们为以往的研究结果和联系建立了理论基础,并为未来HCI在危险识别中的改进提出了具体建议。此外,我们还分析了基于可穿戴振动系统和上下文感知导航的两个与hr - hci领域相关的案例研究。
{"title":"CHR vs. Human-Computer Interaction Design for Emerging Technologies: Two Case Studies","authors":"Sharefa Murad, A. Qusef, Muhanna A. Muhanna","doi":"10.1155/2023/8710638","DOIUrl":"https://doi.org/10.1155/2023/8710638","url":null,"abstract":"Recent years have seen a surge in interest in the multifaceted topic of human-computer interaction (HCI). Since the advent of the Fourth Industrial Revolution, the significance of human-computer interaction in the field of safety risk management has only grown. There has not been a lot of focus on developing human-computer interaction for identifying potential hazards in buildings. After conducting a comprehensive literature review, we developed a study framework for the use of human-computer interaction in the identification of construction-related hazards (CHR-HCI). Future studies will focus on the intersection of computer vision, VR, and ergonomics. In this research, we have built a theoretical foundation for past studies’ findings and connections and offered concrete recommendations for the improvement of HCI in danger identification in the future. Moreover, we analyzed two cases studies related to the domain of CHR-HCI in terms of wearable vibration-based systems and context aware navigation.","PeriodicalId":192934,"journal":{"name":"Adv. Hum. Comput. Interact.","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124634751","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
TSHD: Topic Segmentation Based on Headings Detection (Case Study: Resumes) TSHD:基于标题检测的话题分割(以简历为例)
Pub Date : 2023-02-11 DOI: 10.1155/2023/6044007
Majd E. Tannous, Wassim Ramadan, Mohanad A. Rajab
Many unstructured documents contain segments with specific topics. Extracting these segments and identifying their topics helps to access the required information directly. This can improve the quality of many NLP applications such as information extraction, information retrieval, summarization, and question answering. Resumes (CVs) are unstructured documents that have diverse formats. They contain various segments such as personal information, experience, and education. Manually processing resumes to find the most suitable candidates for a particular job is a difficult task. Due to the increased amount of data, it has become very necessary to manipulate resumes by computer to save time and effort. This research presents a new algorithm named TSHD for topic segmentation based on headings detection. We apply the algorithm to extract resume segments and identify their topics. The proposed TSHD algorithm is accurate and addresses many weaknesses in previous studies. Evaluation results show a very high F1 score (about 96%) and a very low segmentation error (about 2%). The algorithm can be easily adapted to deal with other textual domains that contain headings in their segments.
许多非结构化文档包含具有特定主题的片段。提取这些片段并确定它们的主题有助于直接访问所需的信息。这可以提高许多NLP应用程序的质量,如信息提取、信息检索、摘要和问题回答。简历是一种格式多样的非结构化文档。它们包含个人信息、经历和教育等各个部分。手动处理简历以找到最适合特定职位的候选人是一项艰巨的任务。由于数据量的增加,为了节省时间和精力,用电脑处理简历变得非常必要。本文提出了一种新的基于标题检测的主题分割算法TSHD。我们应用该算法提取简历片段并识别其主题。提出的TSHD算法精度高,解决了以往研究中的许多不足。评价结果表明,该方法的F1分数很高(约96%),分割误差很低(约2%)。该算法可以很容易地适应于处理在其分段中包含标题的其他文本域。
{"title":"TSHD: Topic Segmentation Based on Headings Detection (Case Study: Resumes)","authors":"Majd E. Tannous, Wassim Ramadan, Mohanad A. Rajab","doi":"10.1155/2023/6044007","DOIUrl":"https://doi.org/10.1155/2023/6044007","url":null,"abstract":"Many unstructured documents contain segments with specific topics. Extracting these segments and identifying their topics helps to access the required information directly. This can improve the quality of many NLP applications such as information extraction, information retrieval, summarization, and question answering. Resumes (CVs) are unstructured documents that have diverse formats. They contain various segments such as personal information, experience, and education. Manually processing resumes to find the most suitable candidates for a particular job is a difficult task. Due to the increased amount of data, it has become very necessary to manipulate resumes by computer to save time and effort. This research presents a new algorithm named TSHD for topic segmentation based on headings detection. We apply the algorithm to extract resume segments and identify their topics. The proposed TSHD algorithm is accurate and addresses many weaknesses in previous studies. Evaluation results show a very high F1 score (about 96%) and a very low segmentation error (about 2%). The algorithm can be easily adapted to deal with other textual domains that contain headings in their segments.","PeriodicalId":192934,"journal":{"name":"Adv. Hum. Comput. Interact.","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-02-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129415618","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Advances in the Cognitive Management of Chronic Pain in Children through the Use of Virtual Reality Combined with Binaural Beats: A Pilot Study 虚拟现实结合双耳节拍在儿童慢性疼痛认知管理中的应用进展:一项试点研究
Pub Date : 2022-12-31 DOI: 10.1155/2022/2495182
L. Riera, Sebastià Verger, P. Montoya, Francisco J. Perales López
Chronic pain affects the quality of life of those affected. The need to investigate alternative and complementary methods to the pharmacological one to alleviate chronic pain is evident, so virtual reality and binaural tones have become a topic of interest in this field in recent years. This study aims to analyze the contributions of the combination of these two techniques in pediatric patients with chronic pain. For this, data on psychophysiological responses (heart rate and galvanic skin response) and pain perception are collected during and after interaction with this technology using a mixed pre- and posttest experimental methodology. The physiological data and answers in the Pediatric Pain Questionnaire (PPQ) have been collected in a sample of n = 13 healthy participants and n = 9 pediatric patients with chronic pain. The results show a significant difference between baseline and after applying virtual reality and binaural beats, md = 1.205 (t = 3.32; p < 0.05 ). There are great effects on the perception of chronic pain if virtual reality and binaural beats are combined, even greater than with virtual reality alone, making this combination of technologies a very useful tool to be exploited for the management of chronic pain in pediatric patients with rheumatic diseases.
慢性疼痛会影响患者的生活质量。研究替代和补充药物的方法来缓解慢性疼痛的必要性是显而易见的,因此虚拟现实和双耳音调近年来已成为该领域的研究热点。本研究旨在分析这两种技术在儿科慢性疼痛患者中的作用。为此,使用混合测试前和测试后实验方法,在与该技术交互期间和之后收集心理生理反应(心率和皮肤电反应)和疼痛感知的数据。本研究收集了13名健康儿童和9名慢性疼痛儿童患者的儿童疼痛问卷(PPQ)生理数据和答案。结果显示,基线与应用虚拟现实和双耳节拍后,md = 1.205 (t = 3.32;P < 0.05)。如果虚拟现实和双耳节拍相结合,对慢性疼痛的感知有很大的影响,甚至比单独使用虚拟现实更大,这使得这种技术的结合成为一个非常有用的工具,可以用于治疗患有风湿性疾病的儿科患者的慢性疼痛。
{"title":"Advances in the Cognitive Management of Chronic Pain in Children through the Use of Virtual Reality Combined with Binaural Beats: A Pilot Study","authors":"L. Riera, Sebastià Verger, P. Montoya, Francisco J. Perales López","doi":"10.1155/2022/2495182","DOIUrl":"https://doi.org/10.1155/2022/2495182","url":null,"abstract":"Chronic pain affects the quality of life of those affected. The need to investigate alternative and complementary methods to the pharmacological one to alleviate chronic pain is evident, so virtual reality and binaural tones have become a topic of interest in this field in recent years. This study aims to analyze the contributions of the combination of these two techniques in pediatric patients with chronic pain. For this, data on psychophysiological responses (heart rate and galvanic skin response) and pain perception are collected during and after interaction with this technology using a mixed pre- and posttest experimental methodology. The physiological data and answers in the Pediatric Pain Questionnaire (PPQ) have been collected in a sample of n = 13 healthy participants and n = 9 pediatric patients with chronic pain. The results show a significant difference between baseline and after applying virtual reality and binaural beats, md = 1.205 (t = 3.32; \u0000 \u0000 p\u0000 <\u0000 0.05\u0000 \u0000 ). There are great effects on the perception of chronic pain if virtual reality and binaural beats are combined, even greater than with virtual reality alone, making this combination of technologies a very useful tool to be exploited for the management of chronic pain in pediatric patients with rheumatic diseases.","PeriodicalId":192934,"journal":{"name":"Adv. Hum. Comput. Interact.","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123934422","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Review of the Advancement in Speech Emotion Recognition for Indo-Aryan and Dravidian Languages 印度雅利安语和德拉威语语音情感识别研究进展综述
Pub Date : 2022-12-01 DOI: 10.1155/2022/9602429
Syeda Tamanna Alam Monisha, S. Sultana
Speech emotion recognition (SER) has grown to be one of the most trending research topics in computational linguistics in the last two decades. Speech being the primary communication medium, understanding the emotional state of humans from speech and responding accordingly have made the speech emotion recognition system an essential part of the human-computer interaction (HCI) field. Although there are a few review works carried out for SER, none of them discusses the development of SER system for the Indo-Aryan or Dravidian language families. This paper focuses on some studies carried out for the development of an automatic SER system for Indo-Aryan and Dravidian languages. Besides, it presents a brief study of the prominent databases available for SER experiments. Some remarkable research works on the identification of emotion from the speech signal in the last two decades have also been discussed in this paper.
语音情感识别(SER)是近二十年来计算语言学研究的热点之一。语音作为主要的交流媒介,从语音中了解人的情绪状态并做出相应的反应,使得语音情绪识别系统成为人机交互领域的重要组成部分。虽然对SER进行了一些回顾工作,但没有一个讨论印度雅利安语或德拉威语系SER系统的发展。本文重点介绍了为开发印度雅利安语和德拉威语的自动SER系统而进行的一些研究。此外,本文还简要介绍了可用于SER实验的主要数据库。本文还讨论了近二十年来在语音信号情感识别方面的一些杰出研究成果。
{"title":"A Review of the Advancement in Speech Emotion Recognition for Indo-Aryan and Dravidian Languages","authors":"Syeda Tamanna Alam Monisha, S. Sultana","doi":"10.1155/2022/9602429","DOIUrl":"https://doi.org/10.1155/2022/9602429","url":null,"abstract":"Speech emotion recognition (SER) has grown to be one of the most trending research topics in computational linguistics in the last two decades. Speech being the primary communication medium, understanding the emotional state of humans from speech and responding accordingly have made the speech emotion recognition system an essential part of the human-computer interaction (HCI) field. Although there are a few review works carried out for SER, none of them discusses the development of SER system for the Indo-Aryan or Dravidian language families. This paper focuses on some studies carried out for the development of an automatic SER system for Indo-Aryan and Dravidian languages. Besides, it presents a brief study of the prominent databases available for SER experiments. Some remarkable research works on the identification of emotion from the speech signal in the last two decades have also been discussed in this paper.","PeriodicalId":192934,"journal":{"name":"Adv. Hum. Comput. Interact.","volume":"99 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117235188","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Effect of Social Actors Perceived Agency on Social Presence in Computer-Mediated Communication 社会行动者感知代理对计算机媒介交际中社会存在的影响
Pub Date : 2022-10-22 DOI: 10.1155/2022/9632119
K. Poinsot, G. Gorisse, O. Christmann, Sylvain Fleury, S. Richir
Nowadays, both text-based and 3D online applications rely on conversational agents and autonomous characters to interact with users. Previous experiments demonstrated that perceived agency, that is to say, one’s belief in interacting with a computer- or another human-controlled entity could impact social interaction. At present, theories and results still diverge and no consensus has been reached. Therefore, we developed an experiment to investigate the effect of perceived agency and emotional communication on social presence. Participants were told that they would play an online image recognition game against a computer- or a human-controlled opponent. In both cases, participants faced a computer-controlled opponent designed to provide a challenging yet balanced competitive experience. Depending on the experimental conditions, participants were able to communicate with their opponents using emoticons between the game rounds. Results demonstrate a significant main effect of emotional communication on the three dimensions of social presence we considered in this experiment. An interaction effect between perceived agency and emotional communication was observed in copresence, another core dimension of social presence. The impact of emotional communication on participants’ sense of copresence depends on the perceived agency of the opponent. A significant increase was observed for participants facing a computer-controlled opponent when emotional communication was allowed. The sense of copresence was even higher when they were facing a computer-controlled opponent rather than a presumed human-controlled one. These results are discussed with regard to theories of social interaction in computer-mediated communication.
如今,基于文本和3D的在线应用程序都依赖于会话代理和自主字符与用户交互。先前的实验表明,感知代理,也就是说,一个人在与计算机或另一个人类控制的实体互动时的信念,可能会影响社会互动。目前,理论和结果仍存在分歧,尚未达成共识。因此,我们开展了一项实验来探讨感知代理和情绪沟通对社会存在的影响。参与者被告知,他们将与计算机或人类控制的对手进行在线图像识别游戏。在这两种情况下,参与者都面对一个电脑控制的对手,旨在提供具有挑战性但又平衡的竞争体验。根据实验条件,参与者能够在游戏回合之间使用表情符号与对手交流。结果表明,情绪沟通对社会存在的三个维度有显著的主效应。在社会在场的另一个核心维度——共同在场中,感知代理与情绪沟通之间存在交互作用。情绪沟通对参与者在场感的影响取决于对手的感知代理。当允许进行情感交流时,面对电脑控制的对手的参与者的得分显著增加。当他们面对的是电脑控制的对手,而不是假定的人类控制的对手时,他们的临场感甚至更高。这些结果讨论了社会互动的理论在计算机媒介的沟通。
{"title":"Effect of Social Actors Perceived Agency on Social Presence in Computer-Mediated Communication","authors":"K. Poinsot, G. Gorisse, O. Christmann, Sylvain Fleury, S. Richir","doi":"10.1155/2022/9632119","DOIUrl":"https://doi.org/10.1155/2022/9632119","url":null,"abstract":"Nowadays, both text-based and 3D online applications rely on conversational agents and autonomous characters to interact with users. Previous experiments demonstrated that perceived agency, that is to say, one’s belief in interacting with a computer- or another human-controlled entity could impact social interaction. At present, theories and results still diverge and no consensus has been reached. Therefore, we developed an experiment to investigate the effect of perceived agency and emotional communication on social presence. Participants were told that they would play an online image recognition game against a computer- or a human-controlled opponent. In both cases, participants faced a computer-controlled opponent designed to provide a challenging yet balanced competitive experience. Depending on the experimental conditions, participants were able to communicate with their opponents using emoticons between the game rounds. Results demonstrate a significant main effect of emotional communication on the three dimensions of social presence we considered in this experiment. An interaction effect between perceived agency and emotional communication was observed in copresence, another core dimension of social presence. The impact of emotional communication on participants’ sense of copresence depends on the perceived agency of the opponent. A significant increase was observed for participants facing a computer-controlled opponent when emotional communication was allowed. The sense of copresence was even higher when they were facing a computer-controlled opponent rather than a presumed human-controlled one. These results are discussed with regard to theories of social interaction in computer-mediated communication.","PeriodicalId":192934,"journal":{"name":"Adv. Hum. Comput. Interact.","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132764205","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Digital Mental Health Intervention for Children and Parents Using a User-Centred Design 使用以用户为中心的设计为儿童和家长提供数字心理健康干预
Pub Date : 2022-10-07 DOI: 10.1155/2022/4322177
A. Mahlous, Bersan Okkali
The number of children with mental health problems is ever-growing; as a result, nearly 850,000 children in the UK are believed to have clinically significant problems, and only a quarter show evidence of mental illness. Family members often have a hard time dealing with children with mental health problems. As a result, digital mental health interventions are becoming popular for people seeking professional mental health services. Previous studies in this area have also shown that parents who are divorced or working away from home struggle to maintain contact with their children. This lack of communication between the parents and their children can worsen the children’s mental health conditions and prevent early diagnosis. Human-centred design thinking is applied step by step in this paper to provide an intuitive understanding of the design process. Five stages of the design thinking process were examined to follow a correct path. The results were promising, and the feedback received assured that the product helps parents to better monitor their children’s mental health and provides support when needed. The design thinking process was followed in concordance with the user needs identified from previous studies in this area, which led to a working solution that benefits both parents and children in tackling these problems.
有精神健康问题的儿童人数不断增加;结果,英国近85万名儿童被认为有临床显著问题,只有四分之一的儿童表现出精神疾病的迹象。家庭成员通常很难处理有心理健康问题的孩子。因此,对于寻求专业心理健康服务的人来说,数字心理健康干预措施正变得越来越受欢迎。此前在这一领域的研究也表明,离婚或在外工作的父母很难与孩子保持联系。父母和孩子之间缺乏沟通会恶化孩子的心理健康状况,阻碍早期诊断。以人为本的设计思维在本文中一步一步地应用,以提供对设计过程的直观理解。研究了设计思维过程的五个阶段,以遵循正确的路径。结果是有希望的,收到的反馈确信该产品可以帮助父母更好地监控孩子的心理健康,并在需要时提供支持。设计思维过程遵循了从该领域先前的研究中确定的用户需求,这导致了一个有效的解决方案,在解决这些问题时,父母和孩子都受益。
{"title":"A Digital Mental Health Intervention for Children and Parents Using a User-Centred Design","authors":"A. Mahlous, Bersan Okkali","doi":"10.1155/2022/4322177","DOIUrl":"https://doi.org/10.1155/2022/4322177","url":null,"abstract":"The number of children with mental health problems is ever-growing; as a result, nearly 850,000 children in the UK are believed to have clinically significant problems, and only a quarter show evidence of mental illness. Family members often have a hard time dealing with children with mental health problems. As a result, digital mental health interventions are becoming popular for people seeking professional mental health services. Previous studies in this area have also shown that parents who are divorced or working away from home struggle to maintain contact with their children. This lack of communication between the parents and their children can worsen the children’s mental health conditions and prevent early diagnosis. Human-centred design thinking is applied step by step in this paper to provide an intuitive understanding of the design process. Five stages of the design thinking process were examined to follow a correct path. The results were promising, and the feedback received assured that the product helps parents to better monitor their children’s mental health and provides support when needed. The design thinking process was followed in concordance with the user needs identified from previous studies in this area, which led to a working solution that benefits both parents and children in tackling these problems.","PeriodicalId":192934,"journal":{"name":"Adv. Hum. Comput. Interact.","volume":"89 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115758002","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Generic Approach towards Amharic Sign Language Recognition 阿姆哈拉语手语识别的通用方法
Pub Date : 2022-09-22 DOI: 10.1155/2022/1112169
Netsanet Yigzaw, M. Meshesha, Chala Diriba
In the day-to-day life of communities, good communication channels are crucial for mutual understanding. The hearing-impaired community uses sign language, which is a visual and gestural language. In terms of orientation and expression, it is separate from written and spoken languages. Despite the fact that sign language is an excellent platform for communication among hearing-impaired persons, it has created a communication barrier between hearing-impaired and non-disabled people. To address this issue, researchers have proposed sign language to text translation systems for English and other European languages as a solution. The goal of this research is to design and develop an Amharic digital text converter system using Ethiopian sign language. The proposed system was created with the help of two key deep learning algorithms: a pretrained deep learning model and a Long Short-Term Memory (LSTM). The LSTM was used to extract sequence information from a sequence of image frames of a specific sign language, while the pretrained deep learning model was used to extract features from single frame images. The dataset used to train the algorithms was gathered in video format from Addis Ababa University. Prior to feeding the obtained dataset to the deep learning models, data preprocessing activities such as cleaning and video to image frame segmentation were conducted. The system was trained, validated, and tested using 80%, 10%, and 10% of the 2475 images created during the preprocessing step. Two pretrained deep learning models, EfficientNetB0 and ResNet50, were used in this investigation, and they attained an accuracy of 72.79%. In terms of precision and f1-score, ResNet50 outperformed EfficientNetB0. For the proposed system, a graphical user interface prototype was created, and the best performing model was chosen and implemented. The proposed system can be utilized as a starting point for other researchers to improve upon, based on the outcomes of the experiment. More high-quality training datasets and high-performance training machines, such as GPU-enabled computers, can be added to the system to improve it.
在社区的日常生活中,良好的沟通渠道对相互理解至关重要。听障群体使用手语,这是一种视觉和手势语言。就取向和表达而言,它与书面语和口语是分开的。尽管手语是听障人士之间良好的交流平台,但它在听障人士和非听障人士之间造成了沟通障碍。为了解决这个问题,研究人员提出了英语和其他欧洲语言的手语文本翻译系统作为解决方案。本研究的目的是设计和开发一个使用埃塞俄比亚手语的阿姆哈拉语数字文本转换系统。该系统是在两个关键的深度学习算法的帮助下创建的:一个预训练的深度学习模型和一个长短期记忆(LSTM)。使用LSTM从特定的图像帧序列中提取序列信息,而使用预训练的深度学习模型从单帧图像中提取特征。用于训练算法的数据集以视频格式从亚的斯亚贝巴大学收集。在将获得的数据集输入深度学习模型之前,进行数据预处理活动,如清洗和视频到图像帧分割。使用预处理步骤中创建的2475张图像中的80%、10%和10%对系统进行训练、验证和测试。在本次调查中使用了两个预训练的深度学习模型,有效率netb0和ResNet50,它们的准确率达到了72.79%。在精度和f1-score方面,ResNet50优于EfficientNetB0。针对所提出的系统,创建了图形用户界面原型,选择并实现了性能最佳的模型。所提出的系统可以作为其他研究人员根据实验结果进行改进的起点。可以将更多高质量的训练数据集和高性能的训练机器(如支持gpu的计算机)添加到系统中以改进它。
{"title":"A Generic Approach towards Amharic Sign Language Recognition","authors":"Netsanet Yigzaw, M. Meshesha, Chala Diriba","doi":"10.1155/2022/1112169","DOIUrl":"https://doi.org/10.1155/2022/1112169","url":null,"abstract":"In the day-to-day life of communities, good communication channels are crucial for mutual understanding. The hearing-impaired community uses sign language, which is a visual and gestural language. In terms of orientation and expression, it is separate from written and spoken languages. Despite the fact that sign language is an excellent platform for communication among hearing-impaired persons, it has created a communication barrier between hearing-impaired and non-disabled people. To address this issue, researchers have proposed sign language to text translation systems for English and other European languages as a solution. The goal of this research is to design and develop an Amharic digital text converter system using Ethiopian sign language. The proposed system was created with the help of two key deep learning algorithms: a pretrained deep learning model and a Long Short-Term Memory (LSTM). The LSTM was used to extract sequence information from a sequence of image frames of a specific sign language, while the pretrained deep learning model was used to extract features from single frame images. The dataset used to train the algorithms was gathered in video format from Addis Ababa University. Prior to feeding the obtained dataset to the deep learning models, data preprocessing activities such as cleaning and video to image frame segmentation were conducted. The system was trained, validated, and tested using 80%, 10%, and 10% of the 2475 images created during the preprocessing step. Two pretrained deep learning models, EfficientNetB0 and ResNet50, were used in this investigation, and they attained an accuracy of 72.79%. In terms of precision and f1-score, ResNet50 outperformed EfficientNetB0. For the proposed system, a graphical user interface prototype was created, and the best performing model was chosen and implemented. The proposed system can be utilized as a starting point for other researchers to improve upon, based on the outcomes of the experiment. More high-quality training datasets and high-performance training machines, such as GPU-enabled computers, can be added to the system to improve it.","PeriodicalId":192934,"journal":{"name":"Adv. Hum. Comput. Interact.","volume":"05 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131039835","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Diabetes Mellitus Disease Prediction Using Machine Learning Classifiers with Oversampling and Feature Augmentation 用过采样和特征增强的机器学习分类器预测糖尿病疾病
Pub Date : 2022-09-19 DOI: 10.2139/ssrn.4060653
B. Ahamed, Meenakshi S. Arya, V. O. Nancy
The technical improvements in healthcare sector today have given rise to many new inventions in the field of artificial intelligence. Patterns for disease identification are carried out, and the onset of prediction of many diseases is detected. Diseases include diabetes mellitus disease, fatal heart diseases, and symptomatic cancer. There are many algorithms that have played a critical role in the prediction of diseases. This paper proposes an ML based approach for diabetes mellitus disease prediction. For diabetes prediction, many ML algorithms are compared and used in the proposed work, and finally the three ML classifiers providing the highest accuracy are determined: RF, GBM, and LGBM. The accuracy of prediction is obtained using two types of datasets. They are Pima Indians dataset and a curated dataset. The ML classifiers LGBM, GB, and RF are used to build a predictive model, and the accuracy of each classifier is noted and compared. In addition to the generalized prediction mechanism, the data augmentation technique is also used, and the final accuracy of prediction is obtained for the classifiers LGBM, GB, and RF. A comparative study and demonstration between augmentation and non-augmentation are also discussed for the two datasets used in order to further improve the performance accuracy for predicting diabetes disease.
今天医疗保健领域的技术进步已经在人工智能领域产生了许多新的发明。开展了疾病识别模式,并发现了许多疾病的发病预测。疾病包括糖尿病、致命的心脏病和有症状的癌症。有许多算法在疾病预测中发挥了关键作用。提出了一种基于机器学习的糖尿病疾病预测方法。对于糖尿病预测,我们比较并使用了许多ML算法,最后确定了具有最高准确率的三种ML分类器:RF、GBM和LGBM。使用两种类型的数据集来获得预测的准确性。它们是皮马印第安人数据集和一个策划数据集。使用ML分类器LGBM、GB和RF构建预测模型,并对每个分类器的准确率进行记录和比较。除了广义预测机制外,还采用了数据增强技术,对LGBM、GB和RF分类器进行了最终的预测精度。为了进一步提高预测糖尿病疾病的性能准确性,本文还对两种数据集进行了增强和非增强的对比研究和论证。
{"title":"Diabetes Mellitus Disease Prediction Using Machine Learning Classifiers with Oversampling and Feature Augmentation","authors":"B. Ahamed, Meenakshi S. Arya, V. O. Nancy","doi":"10.2139/ssrn.4060653","DOIUrl":"https://doi.org/10.2139/ssrn.4060653","url":null,"abstract":"The technical improvements in healthcare sector today have given rise to many new inventions in the field of artificial intelligence. Patterns for disease identification are carried out, and the onset of prediction of many diseases is detected. Diseases include diabetes mellitus disease, fatal heart diseases, and symptomatic cancer. There are many algorithms that have played a critical role in the prediction of diseases. This paper proposes an ML based approach for diabetes mellitus disease prediction. For diabetes prediction, many ML algorithms are compared and used in the proposed work, and finally the three ML classifiers providing the highest accuracy are determined: RF, GBM, and LGBM. The accuracy of prediction is obtained using two types of datasets. They are Pima Indians dataset and a curated dataset. The ML classifiers LGBM, GB, and RF are used to build a predictive model, and the accuracy of each classifier is noted and compared. In addition to the generalized prediction mechanism, the data augmentation technique is also used, and the final accuracy of prediction is obtained for the classifiers LGBM, GB, and RF. A comparative study and demonstration between augmentation and non-augmentation are also discussed for the two datasets used in order to further improve the performance accuracy for predicting diabetes disease.","PeriodicalId":192934,"journal":{"name":"Adv. Hum. Comput. Interact.","volume":"2022 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130904153","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
Adv. Hum. Comput. Interact.
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1