首页 > 最新文献

Proceedings of the 2020 International Conference on Multimodal Interaction最新文献

英文 中文
Examining the Link between Children's Cognitive Development and Touchscreen Interaction Patterns 儿童认知发展与触屏互动模式的关系研究
Pub Date : 2020-10-21 DOI: 10.1145/3382507.3418841
Ziyang Chen, Yu-peng Chen, Alex Shaw, Aishat Aloba, Pavlo D. Antonenko, J. Ruiz, Lisa Anthony
It is well established that children's touch and gesture interactions on touchscreen devices are different from those of adults, with much prior work showing that children's input is recognized more poorly than adults? input. In addition, researchers have shown that recognition of touchscreen input is poorest for young children and improves for older children when simply considering their age; however, individual differences in cognitive and motor development could also affect children's input. An understanding of how cognitive and motor skill influence touchscreen interactions, as opposed to only coarser measurements like age and grade level, could help in developing personalized and tailored touchscreen interfaces for each child. To investigate how cognitive and motor development may be related to children's touchscreen interactions, we conducted a study of 28 participants ages 4 to 7 that included validated assessments of the children's motor and cognitive skills as well as typical touchscreen target acquisition and gesture tasks. We correlated participants? touchscreen behaviors to their cognitive development level, including both fine motor skills and executive function. We compare our analysis of touchscreen interactions based on cognitive and motor development to prior work based on children's age. We show that all four factors (age, grade level, motor skill, and executive function) show similar correlations with target miss rates and gesture recognition rates. Thus, we conclude that age and grade level are sufficiently sensitive when considering children's touchscreen behaviors.
众所周知,儿童在触屏设备上的触摸和手势互动与成人不同,许多先前的研究表明,儿童的输入比成人更容易被识别。输入。此外,研究人员还表明,幼儿对触摸屏输入的识别能力最差,而年龄较大的儿童在考虑其年龄时,对触摸屏输入的识别能力有所提高;然而,认知和运动发展的个体差异也可能影响儿童的输入。了解认知和运动技能是如何影响触屏互动的,而不是仅仅像年龄和年级水平这样的粗略测量,可以帮助为每个孩子开发个性化和量身定制的触屏界面。为了研究认知和运动发展与儿童触屏互动的关系,我们对28名4至7岁的儿童进行了一项研究,包括对儿童的运动和认知技能以及典型的触屏目标获取和手势任务的有效评估。我们把参与者联系起来?触屏行为的认知发展水平,包括精细运动技能和执行功能。我们将基于认知和运动发展的触摸屏互动分析与先前基于儿童年龄的研究进行了比较。我们发现,所有四个因素(年龄、年级水平、运动技能和执行功能)都与目标失准率和手势识别率有相似的相关性。因此,我们得出结论,在考虑儿童触屏行为时,年龄和年级是足够敏感的。
{"title":"Examining the Link between Children's Cognitive Development and Touchscreen Interaction Patterns","authors":"Ziyang Chen, Yu-peng Chen, Alex Shaw, Aishat Aloba, Pavlo D. Antonenko, J. Ruiz, Lisa Anthony","doi":"10.1145/3382507.3418841","DOIUrl":"https://doi.org/10.1145/3382507.3418841","url":null,"abstract":"It is well established that children's touch and gesture interactions on touchscreen devices are different from those of adults, with much prior work showing that children's input is recognized more poorly than adults? input. In addition, researchers have shown that recognition of touchscreen input is poorest for young children and improves for older children when simply considering their age; however, individual differences in cognitive and motor development could also affect children's input. An understanding of how cognitive and motor skill influence touchscreen interactions, as opposed to only coarser measurements like age and grade level, could help in developing personalized and tailored touchscreen interfaces for each child. To investigate how cognitive and motor development may be related to children's touchscreen interactions, we conducted a study of 28 participants ages 4 to 7 that included validated assessments of the children's motor and cognitive skills as well as typical touchscreen target acquisition and gesture tasks. We correlated participants? touchscreen behaviors to their cognitive development level, including both fine motor skills and executive function. We compare our analysis of touchscreen interactions based on cognitive and motor development to prior work based on children's age. We show that all four factors (age, grade level, motor skill, and executive function) show similar correlations with target miss rates and gesture recognition rates. Thus, we conclude that age and grade level are sufficiently sensitive when considering children's touchscreen behaviors.","PeriodicalId":402394,"journal":{"name":"Proceedings of the 2020 International Conference on Multimodal Interaction","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126863519","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Towards Engagement Recognition of People with Dementia in Care Settings 面向护理机构中痴呆症患者的参与识别
Pub Date : 2020-10-21 DOI: 10.1145/3382507.3418856
Lars Steinert, F. Putze, Dennis Küster, T. Schultz
Roughly 50 million people worldwide are currently suffering from dementia. This number is expected to triple by 2050. Dementia is characterized by a loss of cognitive function and changes in behaviour. This includes memory, language skills, and the ability to focus and pay attention. However, it has been shown that secondary therapy such as the physical, social and cognitive activation of People with Dementia (PwD) has significant positive effects. Activation impacts cognitive functioning and can help prevent the magnification of apathy, boredom, depression, and loneliness associated with dementia. Furthermore, activation can lead to higher perceived quality of life. We follow Cohen's argument that activation stimuli have to produce engagement to take effect and adopt his definition of engagement as "the act of being occupied or involved with an external stimulus".
目前全世界大约有5000万人患有痴呆症。预计到2050年,这一数字将增加两倍。痴呆症的特点是认知功能丧失和行为改变。这包括记忆力、语言技能以及集中注意力的能力。然而,已有研究表明,对痴呆症患者(PwD)的身体、社会和认知激活等辅助治疗具有显著的积极作用。激活会影响认知功能,并有助于防止与痴呆症相关的冷漠、无聊、抑郁和孤独的放大。此外,激活可以带来更高的感知生活质量。我们遵循Cohen的观点,即激活刺激必须产生参与才能生效,并采用他对参与的定义,即“被外部刺激占据或参与的行为”。
{"title":"Towards Engagement Recognition of People with Dementia in Care Settings","authors":"Lars Steinert, F. Putze, Dennis Küster, T. Schultz","doi":"10.1145/3382507.3418856","DOIUrl":"https://doi.org/10.1145/3382507.3418856","url":null,"abstract":"Roughly 50 million people worldwide are currently suffering from dementia. This number is expected to triple by 2050. Dementia is characterized by a loss of cognitive function and changes in behaviour. This includes memory, language skills, and the ability to focus and pay attention. However, it has been shown that secondary therapy such as the physical, social and cognitive activation of People with Dementia (PwD) has significant positive effects. Activation impacts cognitive functioning and can help prevent the magnification of apathy, boredom, depression, and loneliness associated with dementia. Furthermore, activation can lead to higher perceived quality of life. We follow Cohen's argument that activation stimuli have to produce engagement to take effect and adopt his definition of engagement as \"the act of being occupied or involved with an external stimulus\".","PeriodicalId":402394,"journal":{"name":"Proceedings of the 2020 International Conference on Multimodal Interaction","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122669969","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Multimodal Gated Information Fusion for Emotion Recognition from EEG Signals and Facial Behaviors 基于脑电信号和面部行为的多模态门控信息融合
Pub Date : 2020-10-21 DOI: 10.1145/3382507.3418867
Soheil Rayatdoost, D. Rudrauf, M. Soleymani
Emotions associated with neural and behavioral responses are detectable through scalp electroencephalogram (EEG) signals and measures of facial expressions. We propose a multimodal deep representation learning approach for emotion recognition from EEG and facial expression signals. The proposed method involves the joint learning of a unimodal representation aligned with the other modality through cosine similarity and a gated fusion for modality fusion. We evaluated our method on two databases: DAI-EF and MAHNOB-HCI. The results show that our deep representation is able to learn mutual and complementary information between EEG signals and face video, captured by action units, head and eye movements from face videos, in a manner that generalizes across databases. It is able to outperform similar fusion methods for the task at hand.
与神经和行为反应相关的情绪可以通过头皮脑电图(EEG)信号和面部表情测量来检测。我们提出了一种多模态深度表征学习方法,用于从脑电图和面部表情信号中识别情绪。该方法通过余弦相似度对与其他模态对齐的单模态表示进行联合学习,并对模态融合进行门控融合。我们在两个数据库上评估了我们的方法:DAI-EF和MAHNOB-HCI。结果表明,我们的深度表征能够学习脑电图信号和面部视频之间的相互和互补信息,这些信息是由面部视频中的动作单元、头部和眼睛运动捕获的,以一种跨数据库的方式进行概括。对于手头的任务,它能够胜过类似的融合方法。
{"title":"Multimodal Gated Information Fusion for Emotion Recognition from EEG Signals and Facial Behaviors","authors":"Soheil Rayatdoost, D. Rudrauf, M. Soleymani","doi":"10.1145/3382507.3418867","DOIUrl":"https://doi.org/10.1145/3382507.3418867","url":null,"abstract":"Emotions associated with neural and behavioral responses are detectable through scalp electroencephalogram (EEG) signals and measures of facial expressions. We propose a multimodal deep representation learning approach for emotion recognition from EEG and facial expression signals. The proposed method involves the joint learning of a unimodal representation aligned with the other modality through cosine similarity and a gated fusion for modality fusion. We evaluated our method on two databases: DAI-EF and MAHNOB-HCI. The results show that our deep representation is able to learn mutual and complementary information between EEG signals and face video, captured by action units, head and eye movements from face videos, in a manner that generalizes across databases. It is able to outperform similar fusion methods for the task at hand.","PeriodicalId":402394,"journal":{"name":"Proceedings of the 2020 International Conference on Multimodal Interaction","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123950143","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Multimodal Physiological Synchrony as Measure of Attentional Engagement 多模态生理同步作为注意力投入的测量
Pub Date : 2020-10-21 DOI: 10.1145/3382507.3421152
I. Stuldreher
When interested in monitoring attentional engagement, physiological signals can be of great value. A popular approach is to uncover the complex patterns between physiological signals and attentional engagement using supervised learning models, but it is often unclear which physiological measures can best be used in such models and collecting enough training data with a reliable ground-truth to train such model is very challenging. Rather than using physiological responses of individual participants and specific events in a trained model, one can also continuously determine the degree to which physiological measures of multiple individuals uniformly change, often referred to as physiological synchrony. As a directly proportional relation between physiological synchrony in brain activity and attentional engagement has been pointed out in the literature, no trained model is needed to link the two. I aim to create a more robust measure of attentional engagement among groups of individuals by combining electroencephalography (EEG), electrodermal activity (EDA) and heart rate into a multimodal metric of physiological synchrony. I formulate three main research questions in the current research proposal: 1) How do physiological synchrony in physiological measures from the central and peripheral nervous system relate to attentional engagement? 2) Does physiological synchrony reliably reflect shared attentional engagement in real-world use-cases? 3) How can these physiological measures be fused to obtain a multimodal metric of physiological synchrony that outperforms unimodal synchrony?
当对监测注意力投入感兴趣时,生理信号可能很有价值。一种流行的方法是使用监督学习模型来揭示生理信号和注意力投入之间的复杂模式,但通常不清楚哪种生理测量方法可以最好地用于此类模型,并且收集足够的训练数据并使用可靠的基本事实来训练此类模型非常具有挑战性。与其在训练模型中使用个体参与者的生理反应和特定事件,还可以连续确定多个个体的生理测量均匀变化的程度,通常称为生理同步性。由于已有文献指出大脑活动的生理同步性与注意力投入之间存在直接的正比关系,因此不需要经过训练的模型将两者联系起来。我的目标是通过将脑电图(EEG)、皮电活动(EDA)和心率结合到生理同步的多模态度量中,创建一种更强大的个体群体注意力投入的测量方法。在目前的研究计划中,我提出了三个主要的研究问题:1)中枢和周围神经系统生理测量的生理同步性如何与注意参与相关?2)生理同步是否可靠地反映了现实世界用例中的共同注意力投入?3)如何将这些生理测量融合在一起,以获得优于单峰同步的多模态生理同步度量?
{"title":"Multimodal Physiological Synchrony as Measure of Attentional Engagement","authors":"I. Stuldreher","doi":"10.1145/3382507.3421152","DOIUrl":"https://doi.org/10.1145/3382507.3421152","url":null,"abstract":"When interested in monitoring attentional engagement, physiological signals can be of great value. A popular approach is to uncover the complex patterns between physiological signals and attentional engagement using supervised learning models, but it is often unclear which physiological measures can best be used in such models and collecting enough training data with a reliable ground-truth to train such model is very challenging. Rather than using physiological responses of individual participants and specific events in a trained model, one can also continuously determine the degree to which physiological measures of multiple individuals uniformly change, often referred to as physiological synchrony. As a directly proportional relation between physiological synchrony in brain activity and attentional engagement has been pointed out in the literature, no trained model is needed to link the two. I aim to create a more robust measure of attentional engagement among groups of individuals by combining electroencephalography (EEG), electrodermal activity (EDA) and heart rate into a multimodal metric of physiological synchrony. I formulate three main research questions in the current research proposal: 1) How do physiological synchrony in physiological measures from the central and peripheral nervous system relate to attentional engagement? 2) Does physiological synchrony reliably reflect shared attentional engagement in real-world use-cases? 3) How can these physiological measures be fused to obtain a multimodal metric of physiological synchrony that outperforms unimodal synchrony?","PeriodicalId":402394,"journal":{"name":"Proceedings of the 2020 International Conference on Multimodal Interaction","volume":"85 11-12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124328429","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Using Emotions to Complement Multi-Modal Human-Robot Interaction in Urban Search and Rescue Scenarios 基于情感的多模态人机交互在城市搜救场景中的应用
Pub Date : 2020-10-21 DOI: 10.1145/3382507.3418871
S. Akgun, M. Ghafurian, Mark Crowley, K. Dautenhahn
An experiment is presented to investigate whether there is consensus in mapping emotions to messages/situations in urban search and rescue scenarios, where efficiency and effectiveness of interactions are key to success. We studied mappings between 10 specific messages, presented in two different communication styles, reflecting common situations that might happen during search and rescue missions, and the emotions exhibited by robots in those situations. The data was obtained through a Mechanical Turk study with 78 participants. Our findings support the feasibility of using emotions as an additional communication channel to improve multi-modal human-robot interaction for urban search and rescue robots, and suggests that these mappings are robust, i.e. are not affected by the robot's communication style.
本文提出了一项实验,以调查在城市搜索和救援场景中是否存在将情绪映射到信息/情境的共识,其中互动的效率和有效性是成功的关键。我们研究了10条特定信息之间的映射关系,这些信息以两种不同的沟通方式呈现,反映了搜救任务中可能发生的常见情况,以及机器人在这些情况下表现出的情绪。这些数据是通过一项有78名参与者的土耳其机械研究获得的。我们的研究结果支持了将情感作为一种额外的沟通渠道来改善城市搜救机器人多模态人机交互的可行性,并表明这些映射是鲁棒的,即不受机器人沟通风格的影响。
{"title":"Using Emotions to Complement Multi-Modal Human-Robot Interaction in Urban Search and Rescue Scenarios","authors":"S. Akgun, M. Ghafurian, Mark Crowley, K. Dautenhahn","doi":"10.1145/3382507.3418871","DOIUrl":"https://doi.org/10.1145/3382507.3418871","url":null,"abstract":"An experiment is presented to investigate whether there is consensus in mapping emotions to messages/situations in urban search and rescue scenarios, where efficiency and effectiveness of interactions are key to success. We studied mappings between 10 specific messages, presented in two different communication styles, reflecting common situations that might happen during search and rescue missions, and the emotions exhibited by robots in those situations. The data was obtained through a Mechanical Turk study with 78 participants. Our findings support the feasibility of using emotions as an additional communication channel to improve multi-modal human-robot interaction for urban search and rescue robots, and suggests that these mappings are robust, i.e. are not affected by the robot's communication style.","PeriodicalId":402394,"journal":{"name":"Proceedings of the 2020 International Conference on Multimodal Interaction","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115232982","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
The Sensory Interactive Table: Exploring the Social Space of Eating 感官互动桌:探索饮食的社会空间
Pub Date : 2020-10-21 DOI: 10.1145/3382507.3418866
Roelof Anne Jelle de Vries, Juliet A. M. Haarman, Emiel Harmsen, D. Heylen, H. Hermens
Eating is in many ways a social activity. Yet, little is known about the social dimension of eating influencing individual eating habits. Nor do we know much about how to purposefully design for interactions in the social space of eating. This paper presents (1) the journey of exploring the social space of eating by designing an artifact, and (2) the actual artifact designed for the purpose of exploring the interaction dynamics of social eating. The result of this Research through Design journey is the Sensory Interactive Table: an interactive dining table based on explorations of the social space of eating, and a probe to explore the social space of eating further.
吃东西在很多方面都是一种社交活动。然而,人们对饮食影响个人饮食习惯的社会层面知之甚少。我们也不知道如何有目的地设计饮食社交空间中的互动。本文呈现了(1)通过设计一个人工制品来探索饮食的社会空间的旅程,(2)为探索社会饮食的互动动态而设计的实际人工制品。这次设计之旅的成果就是感官互动桌,这是一个基于对饮食社交空间探索的互动餐桌,是对饮食社交空间进一步探索的探索。
{"title":"The Sensory Interactive Table: Exploring the Social Space of Eating","authors":"Roelof Anne Jelle de Vries, Juliet A. M. Haarman, Emiel Harmsen, D. Heylen, H. Hermens","doi":"10.1145/3382507.3418866","DOIUrl":"https://doi.org/10.1145/3382507.3418866","url":null,"abstract":"Eating is in many ways a social activity. Yet, little is known about the social dimension of eating influencing individual eating habits. Nor do we know much about how to purposefully design for interactions in the social space of eating. This paper presents (1) the journey of exploring the social space of eating by designing an artifact, and (2) the actual artifact designed for the purpose of exploring the interaction dynamics of social eating. The result of this Research through Design journey is the Sensory Interactive Table: an interactive dining table based on explorations of the social space of eating, and a probe to explore the social space of eating further.","PeriodicalId":402394,"journal":{"name":"Proceedings of the 2020 International Conference on Multimodal Interaction","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129318523","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Effects of Visual Locomotion and Tactile Stimuli Duration on the Emotional Dimensions of the Cutaneous Rabbit Illusion 视觉运动和触觉刺激持续时间对皮肤兔错觉情绪维度的影响
Pub Date : 2020-10-21 DOI: 10.1145/3382507.3418835
Mounia Ziat, K. Chin, R. Raisamo
In this study, we assessed the emotional dimensions (valence, arousal, and dominance) of the multimodal visual-cutaneous rabbit effect. Simultaneously to the tactile bursts on the forearm, visual silhouettes of saltatorial animals (rabbit, kangaroo, spider, grasshopper, frog, and flea) were projected on the left arm. Additionally, there were two locomotion conditions: taking-off and landing. The results showed that the valence dimension (happy-unhappy) was only affected by the visual stimuli with no effect of the tactile conditions nor the locomotion phases. Arousal (excited-calm) showed a significant difference for the three tactile conditions with an interaction effect with the locomotion condition. Arousal scores were higher when the taking-off condition was associated with the intermediate duration (24 ms) and when the landing condition was associated with either the shortest duration (12 ms) or the longest duration (48 ms). There was no effect for the dominance dimension. Similar to our previous results, the valence dimension seems to be highly affected by visual information reducing any effect of tactile information, while touch can modulate the arousal dimension. This can be beneficial for designing multimodal interfaces for virtual or augmented reality.
在这项研究中,我们评估了多模态视觉-皮肤兔效应的情感维度(效价、觉醒和优势)。与前臂上的触觉爆发同时,左臂上投射了跳跃动物(兔子、袋鼠、蜘蛛、蚱蜢、青蛙和跳蚤)的视觉轮廓。此外,还有两种运动条件:起飞和着陆。结果表明,效价维度(快乐-不快乐)仅受视觉刺激的影响,而不受触觉条件和运动阶段的影响。唤醒(兴奋-平静)在三种触觉条件下表现出显著差异,并与运动条件相互作用。当起飞条件与中间持续时间(24毫秒)相关联时,当着陆条件与最短持续时间(12毫秒)或最长持续时间(48毫秒)相关联时,唤醒分数更高。优势度维度不受影响。与我们之前的结果相似,效价维度似乎受到视觉信息的高度影响,减少了触觉信息的任何影响,而触觉可以调节唤醒维度。这对于设计虚拟或增强现实的多模态界面是有益的。
{"title":"Effects of Visual Locomotion and Tactile Stimuli Duration on the Emotional Dimensions of the Cutaneous Rabbit Illusion","authors":"Mounia Ziat, K. Chin, R. Raisamo","doi":"10.1145/3382507.3418835","DOIUrl":"https://doi.org/10.1145/3382507.3418835","url":null,"abstract":"In this study, we assessed the emotional dimensions (valence, arousal, and dominance) of the multimodal visual-cutaneous rabbit effect. Simultaneously to the tactile bursts on the forearm, visual silhouettes of saltatorial animals (rabbit, kangaroo, spider, grasshopper, frog, and flea) were projected on the left arm. Additionally, there were two locomotion conditions: taking-off and landing. The results showed that the valence dimension (happy-unhappy) was only affected by the visual stimuli with no effect of the tactile conditions nor the locomotion phases. Arousal (excited-calm) showed a significant difference for the three tactile conditions with an interaction effect with the locomotion condition. Arousal scores were higher when the taking-off condition was associated with the intermediate duration (24 ms) and when the landing condition was associated with either the shortest duration (12 ms) or the longest duration (48 ms). There was no effect for the dominance dimension. Similar to our previous results, the valence dimension seems to be highly affected by visual information reducing any effect of tactile information, while touch can modulate the arousal dimension. This can be beneficial for designing multimodal interfaces for virtual or augmented reality.","PeriodicalId":402394,"journal":{"name":"Proceedings of the 2020 International Conference on Multimodal Interaction","volume":"598 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125621107","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Neural Architecture for Detecting User Confusion in Eye-tracking Data 一种检测眼动数据中用户混淆的神经结构
Pub Date : 2020-10-21 DOI: 10.1145/3382507.3418828
Shane D. V. Sims, C. Conati
Encouraged by the success of deep learning in a variety of domains, we investigate the effectiveness of a novel application of such methods for detecting user confusion with eye-tracking data. We introduce an architecture that uses RNN and CNN sub-models in parallel, to take advantage of the temporal and visuospatial aspects of our data. Experiments with a dataset of user interactions with the ValueChart visualization tool show that our model outperforms an existing model based on a Random Forest classifier, resulting in a 22% improvement in combined confused & not confused class accuracies.
受深度学习在各个领域取得成功的鼓舞,我们研究了这种方法在使用眼动追踪数据检测用户混淆的新应用的有效性。我们引入了一个并行使用RNN和CNN子模型的架构,以利用我们数据的时间和视觉空间方面。使用ValueChart可视化工具的用户交互数据集进行的实验表明,我们的模型优于基于随机森林分类器的现有模型,导致混淆和非混淆类别的组合准确率提高22%。
{"title":"A Neural Architecture for Detecting User Confusion in Eye-tracking Data","authors":"Shane D. V. Sims, C. Conati","doi":"10.1145/3382507.3418828","DOIUrl":"https://doi.org/10.1145/3382507.3418828","url":null,"abstract":"Encouraged by the success of deep learning in a variety of domains, we investigate the effectiveness of a novel application of such methods for detecting user confusion with eye-tracking data. We introduce an architecture that uses RNN and CNN sub-models in parallel, to take advantage of the temporal and visuospatial aspects of our data. Experiments with a dataset of user interactions with the ValueChart visualization tool show that our model outperforms an existing model based on a Random Forest classifier, resulting in a 22% improvement in combined confused & not confused class accuracies.","PeriodicalId":402394,"journal":{"name":"Proceedings of the 2020 International Conference on Multimodal Interaction","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131736173","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
StrategicReading StrategicReading
Pub Date : 2020-10-21 DOI: 10.1145/3382507.3418879
W. Guo, Byeong-Young Cho, Jingtao Wang
Mobile devices are becoming an important platform for reading. However, existing research on mobile reading primarily focuses on low-level metrics such as speed and comprehension. For complex reading tasks involving information seeking and context switching, researchers still rely on verbal reports via think-aloud. We present StrategicReading, an intelligent reading system running on unmodified smartphones, to understand high-level strategic reading behaviors on mobile devices. StrategicReading leverages multimodal behavior sensing and takes advantage of signals from camera-based gaze sensing, kinematic scrolling patterns, and cross-page behavior changes. Through a 40-participant study, we found that gaze patterns, muscle stiffness signals, and reading paths captured by StrategicReading can infer both users' reading strategies and reading performance with high accuracy.
{"title":"StrategicReading","authors":"W. Guo, Byeong-Young Cho, Jingtao Wang","doi":"10.1145/3382507.3418879","DOIUrl":"https://doi.org/10.1145/3382507.3418879","url":null,"abstract":"Mobile devices are becoming an important platform for reading. However, existing research on mobile reading primarily focuses on low-level metrics such as speed and comprehension. For complex reading tasks involving information seeking and context switching, researchers still rely on verbal reports via think-aloud. We present StrategicReading, an intelligent reading system running on unmodified smartphones, to understand high-level strategic reading behaviors on mobile devices. StrategicReading leverages multimodal behavior sensing and takes advantage of signals from camera-based gaze sensing, kinematic scrolling patterns, and cross-page behavior changes. Through a 40-participant study, we found that gaze patterns, muscle stiffness signals, and reading paths captured by StrategicReading can infer both users' reading strategies and reading performance with high accuracy.","PeriodicalId":402394,"journal":{"name":"Proceedings of the 2020 International Conference on Multimodal Interaction","volume":"417 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123271656","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Musical Multimodal Interaction: From Bodies to Ecologies 音乐多模态互动:从身体到生态
Pub Date : 2020-10-21 DOI: 10.1145/3382507.3419444
Atau Tanaka
Musical performance can be thought of in multimodal terms - physical interaction with musical instruments produces sound output, often while the performer is visually reading a score. Digital Musical Instrument (DMI) design merges tenets of HCI and musical instrument practice. Audiovisual performance and other forms of multimedia might benefit from multimodal thinking. This keynote revisits two decades of interactive music practice that has paralleled the development of the field of multimodal interaction research. The BioMuse was an early digital musical instrument system using EMG muscle sensing that was extended by a second mode of sensing, allowing effort and position to be two complementary modalities [1]. The Haptic Wave applied principles of cross-modal information display to create a haptic audio editor enabling visually impaired audio producers to 'feel' audio waveforms they could not see in a graphical user interface [2]. VJ culture extends the idea of music DJs to create audiovisual cultural experiences. AVUIs were a set of creative coding tools that enabled the convergence of performance UI and creative visual output [3]. The Orchestra of Rocks is a continuing collaboration with visual artist Uta Kogelsberger that has manifested itself through physical and virtual forms - allowing multimodality over time [4]. Be it a physical exhibition in a gallery or audio reactive 3D animation on YouTube 360, the multiple modes in which an artwork is articulated support its original conceptual foundations. These four projects situate multimodal interaction at the heart of artistic research.
音乐表演可以用多模态的术语来考虑——与乐器的物理互动产生声音输出,通常是在表演者视觉上阅读乐谱的时候。数字乐器(DMI)设计融合了HCI和乐器实践的原则。视听表演和其他形式的多媒体可能受益于多模态思维。这个主题回顾了二十年来与多模态互动研究领域的发展平行的互动音乐实践。BioMuse是一种早期的数字乐器系统,它使用肌电图肌肉感应,并通过第二种感应模式进行扩展,使力度和位置成为两种互补的模式[1]。Haptic Wave应用了跨模态信息显示的原理来创建一个触觉音频编辑器,使视障音频制作者能够“感觉”到他们在图形用户界面中看不到的音频波形[2]。VJ文化将音乐dj的理念延伸到创造视听文化体验。avui是一套创造性的编码工具,实现了性能UI和创造性视觉输出的融合[3]。《岩石管弦乐队》是与视觉艺术家Uta Kogelsberger的持续合作,通过物理和虚拟形式表现出来,随着时间的推移允许多模态[4]。无论是画廊中的实体展览还是YouTube 360上的音频反应3D动画,艺术作品的多种模式都支持其原始概念基础。这四个项目将多模式互动置于艺术研究的核心。
{"title":"Musical Multimodal Interaction: From Bodies to Ecologies","authors":"Atau Tanaka","doi":"10.1145/3382507.3419444","DOIUrl":"https://doi.org/10.1145/3382507.3419444","url":null,"abstract":"Musical performance can be thought of in multimodal terms - physical interaction with musical instruments produces sound output, often while the performer is visually reading a score. Digital Musical Instrument (DMI) design merges tenets of HCI and musical instrument practice. Audiovisual performance and other forms of multimedia might benefit from multimodal thinking. This keynote revisits two decades of interactive music practice that has paralleled the development of the field of multimodal interaction research. The BioMuse was an early digital musical instrument system using EMG muscle sensing that was extended by a second mode of sensing, allowing effort and position to be two complementary modalities [1]. The Haptic Wave applied principles of cross-modal information display to create a haptic audio editor enabling visually impaired audio producers to 'feel' audio waveforms they could not see in a graphical user interface [2]. VJ culture extends the idea of music DJs to create audiovisual cultural experiences. AVUIs were a set of creative coding tools that enabled the convergence of performance UI and creative visual output [3]. The Orchestra of Rocks is a continuing collaboration with visual artist Uta Kogelsberger that has manifested itself through physical and virtual forms - allowing multimodality over time [4]. Be it a physical exhibition in a gallery or audio reactive 3D animation on YouTube 360, the multiple modes in which an artwork is articulated support its original conceptual foundations. These four projects situate multimodal interaction at the heart of artistic research.","PeriodicalId":402394,"journal":{"name":"Proceedings of the 2020 International Conference on Multimodal Interaction","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116274243","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of the 2020 International Conference on Multimodal Interaction
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1