首页 > 最新文献

Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies最新文献

英文 中文
Model Touch Pointing and Detect Parkinson's Disease via a Mobile Game 通过手机游戏模拟触摸指向并检测帕金森病
Q1 Computer Science Pub Date : 2024-05-13 DOI: 10.1145/3659627
Kaiyan Ling, Hang Zhao, Xiangmin Fan, Xiaohui Niu, Wenchao Yin, Yue Liu, Cui Wang, Xiaojun Bi
Touch pointing is one of the primary interaction actions on mobile devices. In this research, we aim to (1) model touch pointing for people with Parkinson's Disease (PD), and (2) detect PD via touch pointing. We created a mobile game called MoleBuster in which a user performs a sequence of pointing actions. Our study with 40 participants shows that PD participants exhibited distinct pointing behavior. PD participants were much slower and had greater variances in movement time (MT), while their error rate was slightly lower than age-matched non-PD participants, indicating PD participants traded speed for accuracy. The nominal width Finger-Fitts law showed greater fitness than Fitts' law, suggesting this model should be adopted in lieu of Fitts' law to guide mobile interface design for PD users. We also proposed a CNN-Transformer-based neural network model to detect PD. Taking touch pointing data and comfort rating of finger movement as input, this model achieved an AUC of 0.97 and sensitivity of 0.95 in leave-one-user-out cross-validation. Overall, our research contributes models that reveal the temporal and spatial characteristics of touch pointing for PD users, and provide a new method (CNN-Transformer model) and a mobile game (MoleBuster) for convenient PD detection.
触摸指向是移动设备上的主要交互操作之一。在这项研究中,我们的目标是:(1)为帕金森病(PD)患者的触摸指向建模;(2)通过触摸指向检测帕金森病。我们制作了一款名为 "鼹鼠克星"(MoleBuster)的手机游戏,用户可以在游戏中执行一连串的指向操作。我们对 40 名参与者进行的研究表明,帕金森病参与者表现出与众不同的指向行为。患有肢体麻痹症的参与者速度更慢,移动时间(MT)的差异更大,而他们的错误率略低于年龄匹配的非肢体麻痹症参与者,这表明患有肢体麻痹症的参与者以速度换取了准确性。标称宽度的 Finger-Fitts 定律比 Fitts 定律显示出更高的适配性,这表明应采用该模型代替 Fitts 定律来指导针对 PD 用户的移动界面设计。我们还提出了一个基于 CNN 变换器的神经网络模型来检测肢体麻木症。以触摸指向数据和手指运动舒适度评级为输入,该模型在 "留一用户 "交叉验证中取得了 0.97 的 AUC 和 0.95 的灵敏度。总之,我们的研究为揭示肢端麻痹用户触摸指向的时间和空间特征提供了模型,并为便捷地检测肢端麻痹提供了一种新方法(CNN-Transformer 模型)和一款手机游戏(MoleBuster)。
{"title":"Model Touch Pointing and Detect Parkinson's Disease via a Mobile Game","authors":"Kaiyan Ling, Hang Zhao, Xiangmin Fan, Xiaohui Niu, Wenchao Yin, Yue Liu, Cui Wang, Xiaojun Bi","doi":"10.1145/3659627","DOIUrl":"https://doi.org/10.1145/3659627","url":null,"abstract":"Touch pointing is one of the primary interaction actions on mobile devices. In this research, we aim to (1) model touch pointing for people with Parkinson's Disease (PD), and (2) detect PD via touch pointing. We created a mobile game called MoleBuster in which a user performs a sequence of pointing actions. Our study with 40 participants shows that PD participants exhibited distinct pointing behavior. PD participants were much slower and had greater variances in movement time (MT), while their error rate was slightly lower than age-matched non-PD participants, indicating PD participants traded speed for accuracy. The nominal width Finger-Fitts law showed greater fitness than Fitts' law, suggesting this model should be adopted in lieu of Fitts' law to guide mobile interface design for PD users. We also proposed a CNN-Transformer-based neural network model to detect PD. Taking touch pointing data and comfort rating of finger movement as input, this model achieved an AUC of 0.97 and sensitivity of 0.95 in leave-one-user-out cross-validation. Overall, our research contributes models that reveal the temporal and spatial characteristics of touch pointing for PD users, and provide a new method (CNN-Transformer model) and a mobile game (MoleBuster) for convenient PD detection.","PeriodicalId":20553,"journal":{"name":"Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140983604","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Technology which Makes You Think 让你思考的技术
Q1 Computer Science Pub Date : 2024-05-13 DOI: 10.1145/3659615
Meagan B. Loerakker, Jasmin Niess, Paweł W. Woźniak
Reflection is widely regarded as a key design goal for technologies for well-being. Yet, recent research shows that technologies for reflection may have negative consequences, in the form of rumination, i.e. negative thought cycles. Understanding how technologies support thinking about oneself, which can take the form of rumination and reflection, is key for future well-being technologies. To address this research gap, we developed the Reflection, Rumination and Thought in Technology (R2T2) scale. Contrary to past research, R2T2 addresses ways of self-focused thinking beyond reflection. This scale can quantify how a technology supports self-focused thinking and the rumination and reflection aspects of that thinking. We developed the scale through a systematic scale development process. We then evaluated the scale's test-retest reliability along with its concurrent and discriminant validity. R2T2 enables designers and researchers to compare technologies which embrace self-focused thinking and its facets as a design goal.
反思被广泛视为福祉技术的一个关键设计目标。然而,最近的研究表明,反思技术可能会产生负面影响,表现为反刍,即消极的思维循环。了解技术如何支持以反刍和反思为形式的自我思考,是未来福祉技术的关键。为了填补这一研究空白,我们开发了科技中的反思、反刍和思考量表(R2T2)。与以往的研究不同,R2T2 所涉及的是反思之外的自我思考方式。该量表可以量化一项技术如何支持以自我为中心的思考,以及这种思考的反刍和反思方面。我们通过系统化的量表开发流程开发了该量表。然后,我们评估了量表的重测信度、并发效度和判别效度。R2T2 使设计者和研究人员能够比较将自我专注思维及其各个方面作为设计目标的技术。
{"title":"Technology which Makes You Think","authors":"Meagan B. Loerakker, Jasmin Niess, Paweł W. Woźniak","doi":"10.1145/3659615","DOIUrl":"https://doi.org/10.1145/3659615","url":null,"abstract":"Reflection is widely regarded as a key design goal for technologies for well-being. Yet, recent research shows that technologies for reflection may have negative consequences, in the form of rumination, i.e. negative thought cycles. Understanding how technologies support thinking about oneself, which can take the form of rumination and reflection, is key for future well-being technologies. To address this research gap, we developed the Reflection, Rumination and Thought in Technology (R2T2) scale. Contrary to past research, R2T2 addresses ways of self-focused thinking beyond reflection. This scale can quantify how a technology supports self-focused thinking and the rumination and reflection aspects of that thinking. We developed the scale through a systematic scale development process. We then evaluated the scale's test-retest reliability along with its concurrent and discriminant validity. R2T2 enables designers and researchers to compare technologies which embrace self-focused thinking and its facets as a design goal.","PeriodicalId":20553,"journal":{"name":"Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140983199","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The EarSAVAS Dataset EarSAVAS 数据集
Q1 Computer Science Pub Date : 2024-05-13 DOI: 10.1145/3659616
Xiyuxing Zhang, Yuntao Wang, Yuxuan Han, Chen Liang, Ishan Chatterjee, Jiankai Tang, Xin Yi, Shwetak Patel, Yuanchun Shi
Subject-aware vocal activity sensing on wearables, which specifically recognizes and monitors the wearer's distinct vocal activities, is essential in advancing personal health monitoring and enabling context-aware applications. While recent advancements in earables present new opportunities, the absence of relevant datasets and effective methods remains a significant challenge. In this paper, we introduce EarSAVAS, the first publicly available dataset constructed specifically for subject-aware human vocal activity sensing on earables. EarSAVAS encompasses eight distinct vocal activities from both the earphone wearer and bystanders, including synchronous two-channel audio and motion data collected from 42 participants totaling 44.5 hours. Further, we propose EarVAS, a lightweight multi-modal deep learning architecture that enables efficient subject-aware vocal activity recognition on earables. To validate the reliability of EarSAVAS and the efficiency of EarVAS, we implemented two advanced benchmark models. Evaluation results on EarSAVAS reveal EarVAS's effectiveness with an accuracy of 90.84% and a Macro-AUC of 89.03%. Comprehensive ablation experiments were conducted on benchmark models and demonstrated the effectiveness of feedback microphone audio and highlighted the potential value of sensor fusion in subject-aware vocal activity sensing on earables. We hope that the proposed EarSAVAS and benchmark models can inspire other researchers to further explore efficient subject-aware human vocal activity sensing on earables.
可穿戴设备上的主体感知发声活动传感可专门识别和监测穿戴者的独特发声活动,对于推进个人健康监测和实现情境感知应用至关重要。虽然耳戴设备的最新进展带来了新的机遇,但缺乏相关数据集和有效方法仍是一个重大挑战。在本文中,我们将介绍 EarSAVAS,它是首个公开可用的数据集,专门用于在耳机上构建主体感知人类发声活动传感。EarSAVAS 包含耳机佩戴者和旁观者的八种不同的发声活动,其中包括从 42 名参与者处收集的同步双通道音频和运动数据,总时长 44.5 小时。此外,我们还提出了一种轻量级多模态深度学习架构--EarVAS,该架构可在耳机上实现高效的主体感知发声活动识别。为了验证 EarSAVAS 的可靠性和 EarVAS 的效率,我们实施了两个先进的基准模型。对 EarSAVAS 的评估结果显示了 EarVAS 的有效性,其准确率为 90.84%,Macro-AUC 为 89.03%。我们在基准模型上进行了全面的消融实验,证明了反馈麦克风音频的有效性,并强调了传感器融合在耳机主体感知发声活动传感中的潜在价值。我们希望所提出的 EarSAVAS 和基准模型能激励其他研究人员进一步探索在耳机上进行高效的主体感知人类发声活动传感。
{"title":"The EarSAVAS Dataset","authors":"Xiyuxing Zhang, Yuntao Wang, Yuxuan Han, Chen Liang, Ishan Chatterjee, Jiankai Tang, Xin Yi, Shwetak Patel, Yuanchun Shi","doi":"10.1145/3659616","DOIUrl":"https://doi.org/10.1145/3659616","url":null,"abstract":"Subject-aware vocal activity sensing on wearables, which specifically recognizes and monitors the wearer's distinct vocal activities, is essential in advancing personal health monitoring and enabling context-aware applications. While recent advancements in earables present new opportunities, the absence of relevant datasets and effective methods remains a significant challenge. In this paper, we introduce EarSAVAS, the first publicly available dataset constructed specifically for subject-aware human vocal activity sensing on earables. EarSAVAS encompasses eight distinct vocal activities from both the earphone wearer and bystanders, including synchronous two-channel audio and motion data collected from 42 participants totaling 44.5 hours. Further, we propose EarVAS, a lightweight multi-modal deep learning architecture that enables efficient subject-aware vocal activity recognition on earables. To validate the reliability of EarSAVAS and the efficiency of EarVAS, we implemented two advanced benchmark models. Evaluation results on EarSAVAS reveal EarVAS's effectiveness with an accuracy of 90.84% and a Macro-AUC of 89.03%. Comprehensive ablation experiments were conducted on benchmark models and demonstrated the effectiveness of feedback microphone audio and highlighted the potential value of sensor fusion in subject-aware vocal activity sensing on earables. We hope that the proposed EarSAVAS and benchmark models can inspire other researchers to further explore efficient subject-aware human vocal activity sensing on earables.","PeriodicalId":20553,"journal":{"name":"Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140985947","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Talk2Care: An LLM-based Voice Assistant for Communication between Healthcare Providers and Older Adults Talk2Care:基于 LLM 的语音助手,用于医护人员与老年人之间的交流
Q1 Computer Science Pub Date : 2024-05-13 DOI: 10.1145/3659625
Ziqi Yang, Xuhai Xu, Bingsheng Yao, Ethan Rogers, Shao Zhang, Stephen Intille, Nawar Shara, G. Gao, Dakuo Wang
Despite the plethora of telehealth applications to assist home-based older adults and healthcare providers, basic messaging and phone calls are still the most common communication methods, which suffer from limited availability, information loss, and process inefficiencies. One promising solution to facilitate patient-provider communication is to leverage large language models (LLMs) with their powerful natural conversation and summarization capability. However, there is a limited understanding of LLMs' role during the communication. We first conducted two interview studies with both older adults (N=10) and healthcare providers (N=9) to understand their needs and opportunities for LLMs in patient-provider asynchronous communication. Based on the insights, we built an LLM-powered communication system, Talk2Care, and designed interactive components for both groups: (1) For older adults, we leveraged the convenience and accessibility of voice assistants (VAs) and built an LLM-powered conversational interface for effective information collection. (2) For health providers, we built an LLM-based dashboard to summarize and present important health information based on older adults' conversations with the VA. We further conducted two user studies with older adults and providers to evaluate the usability of the system. The results showed that Talk2Care could facilitate the communication process, enrich the health information collected from older adults, and considerably save providers' efforts and time. We envision our work as an initial exploration of LLMs' capability in the intersection of healthcare and interpersonal communication.
尽管有大量的远程医疗应用为在家的老年人和医疗服务提供者提供帮助,但基本的信息传递和电话仍是最常用的沟通方式,它们存在可用性有限、信息丢失和流程效率低下等问题。促进患者与医疗服务提供者沟通的一个有前途的解决方案是利用大型语言模型(LLMs)强大的自然会话和总结能力。然而,人们对 LLM 在沟通过程中的作用了解有限。我们首先对老年人(10 人)和医疗保健提供者(9 人)进行了两次访谈研究,以了解他们在患者与提供者异步交流中对 LLM 的需求和机会。基于这些见解,我们建立了一个由 LLM 驱动的通信系统 Talk2Care,并为这两个群体设计了互动组件:(1)对于老年人,我们利用语音助手(VA)的便利性和可及性,建立了一个由 LLM 驱动的对话界面,以有效收集信息。(2)对于医疗服务提供者,我们建立了一个基于 LLM 的仪表板,根据老年人与 VA 的对话总结并展示重要的健康信息。我们还对老年人和医疗服务提供者进行了两次用户研究,以评估系统的可用性。结果表明,Talk2Care 可以促进沟通过程,丰富从老年人那里收集到的健康信息,并大大节省医疗服务提供者的精力和时间。我们认为,我们的工作是对 LLMs 在医疗保健和人际沟通交叉领域能力的初步探索。
{"title":"Talk2Care: An LLM-based Voice Assistant for Communication between Healthcare Providers and Older Adults","authors":"Ziqi Yang, Xuhai Xu, Bingsheng Yao, Ethan Rogers, Shao Zhang, Stephen Intille, Nawar Shara, G. Gao, Dakuo Wang","doi":"10.1145/3659625","DOIUrl":"https://doi.org/10.1145/3659625","url":null,"abstract":"Despite the plethora of telehealth applications to assist home-based older adults and healthcare providers, basic messaging and phone calls are still the most common communication methods, which suffer from limited availability, information loss, and process inefficiencies. One promising solution to facilitate patient-provider communication is to leverage large language models (LLMs) with their powerful natural conversation and summarization capability. However, there is a limited understanding of LLMs' role during the communication. We first conducted two interview studies with both older adults (N=10) and healthcare providers (N=9) to understand their needs and opportunities for LLMs in patient-provider asynchronous communication. Based on the insights, we built an LLM-powered communication system, Talk2Care, and designed interactive components for both groups: (1) For older adults, we leveraged the convenience and accessibility of voice assistants (VAs) and built an LLM-powered conversational interface for effective information collection. (2) For health providers, we built an LLM-based dashboard to summarize and present important health information based on older adults' conversations with the VA. We further conducted two user studies with older adults and providers to evaluate the usability of the system. The results showed that Talk2Care could facilitate the communication process, enrich the health information collected from older adults, and considerably save providers' efforts and time. We envision our work as an initial exploration of LLMs' capability in the intersection of healthcare and interpersonal communication.","PeriodicalId":20553,"journal":{"name":"Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140982074","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CrowdBot: An Open-Environment Robot Management System for On-Campus Services CrowdBot:用于校内服务的开放环境机器人管理系统
Q1 Computer Science Pub Date : 2024-05-13 DOI: 10.1145/3659601
Yufei Wang, Wenting Zeng, Changzhen Liu, Zhuohan Ye, Jiawei Sun, Junxiang Ji, Zhihan Jiang, Xianyi Yan, Yongyi Wu, Yigao Wang, Dingqi Yang, Leye Wang, Daqing Zhang, Cheng Wang, Longbiao Chen
In contemporary campus environments, the provision of timely and efficient services is increasingly challenging due to limitations in accessibility and the complexity and openness of the environment. Existing service robots, while operational, often struggle with adaptability and dynamic task management, leading to inefficiencies. To overcome these limitations, we introduce CrowdBot, a robot management system that enhances service in campus environments. Our system leverages a hierarchical reinforcement learning-based cloud-edge hybrid scheduling framework (REDIS), for efficient online streaming task assignment and dynamic action scheduling. To verify the REDIS framework, we have developed a digital twin simulation platform, which integrates large language models and hot-swapping technology. This facilitates seamless human-robot interaction, efficient task allocation, and cost-effective execution through the reuse of robot equipment. Our comprehensive simulations corroborate the system's remarkable efficacy, demonstrating significant improvements with a 24.46% reduction in task completion times, a 9.37% decrease in travel distances, and up to a 3% savings in power usage. Additionally, the system achieves a 7.95% increase in the number of tasks completed and a 9.49% reduction in response time. Real-world case studies further affirm CrowdBot's capability to adeptly execute tasks and judiciously recycle resources, thereby offering a smart and viable solution for the streamlined management of campus services.
在当代校园环境中,由于交通不便以及环境的复杂性和开放性,提供及时高效的服务越来越具有挑战性。现有的服务机器人虽然可以运行,但在适应性和动态任务管理方面往往力不从心,导致效率低下。为了克服这些局限性,我们推出了 CrowdBot 机器人管理系统,以增强校园环境中的服务。我们的系统利用基于分层强化学习的云边混合调度框架(REDIS),实现高效的在线流式任务分配和动态行动调度。为了验证 REDIS 框架,我们开发了一个数字孪生模拟平台,该平台集成了大型语言模型和热插拔技术。这有助于实现无缝的人机交互、高效的任务分配以及通过重复使用机器人设备实现经济高效的执行。我们的综合模拟证实了该系统的显著功效,任务完成时间缩短了 24.46%,行进距离缩短了 9.37%,耗电量节省了 3%。此外,该系统完成的任务数量增加了 7.95%,响应时间缩短了 9.49%。现实世界的案例研究进一步肯定了 CrowdBot 熟练执行任务和明智回收资源的能力,从而为简化校园服务管理提供了一个智能而可行的解决方案。
{"title":"CrowdBot: An Open-Environment Robot Management System for On-Campus Services","authors":"Yufei Wang, Wenting Zeng, Changzhen Liu, Zhuohan Ye, Jiawei Sun, Junxiang Ji, Zhihan Jiang, Xianyi Yan, Yongyi Wu, Yigao Wang, Dingqi Yang, Leye Wang, Daqing Zhang, Cheng Wang, Longbiao Chen","doi":"10.1145/3659601","DOIUrl":"https://doi.org/10.1145/3659601","url":null,"abstract":"In contemporary campus environments, the provision of timely and efficient services is increasingly challenging due to limitations in accessibility and the complexity and openness of the environment. Existing service robots, while operational, often struggle with adaptability and dynamic task management, leading to inefficiencies. To overcome these limitations, we introduce CrowdBot, a robot management system that enhances service in campus environments. Our system leverages a hierarchical reinforcement learning-based cloud-edge hybrid scheduling framework (REDIS), for efficient online streaming task assignment and dynamic action scheduling. To verify the REDIS framework, we have developed a digital twin simulation platform, which integrates large language models and hot-swapping technology. This facilitates seamless human-robot interaction, efficient task allocation, and cost-effective execution through the reuse of robot equipment. Our comprehensive simulations corroborate the system's remarkable efficacy, demonstrating significant improvements with a 24.46% reduction in task completion times, a 9.37% decrease in travel distances, and up to a 3% savings in power usage. Additionally, the system achieves a 7.95% increase in the number of tasks completed and a 9.49% reduction in response time. Real-world case studies further affirm CrowdBot's capability to adeptly execute tasks and judiciously recycle resources, thereby offering a smart and viable solution for the streamlined management of campus services.","PeriodicalId":20553,"journal":{"name":"Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140983283","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SwivelTouch: Boosting Touchscreen Input with 3D Finger Rotation Gesture SwivelTouch:利用 3D 手指旋转手势增强触摸屏输入功能
Q1 Computer Science Pub Date : 2024-05-13 DOI: 10.1145/3659584
Chentao Li, Jinyang Yu, Ke He, Jianjiang Feng, Jie Zhou
Today, touchscreens stand as the most prevalent input devices of mobile computing devices (smartphones, tablets, smartwatches). Yet, compared with desktop or laptop computers, the limited shortcut keys and physical buttons on touchscreen devices, coupled with the fat finger problem, often lead to slower and more error-prone input and navigation, especially when dealing with text editing and other complex interaction tasks. We introduce an innovative gesture set based on finger rotations in the yaw, pitch, and roll directions on a touchscreen, diverging significantly from traditional two-dimensional interactions and promising to expand the gesture library. Despite active research in estimation of finger angles, however, the previous work faces substantial challenges, including significant estimation errors and unstable sequential outputs. Variability in user behavior further complicates the isolation of movements to a single rotational axis, leading to accidental disturbances and screen coordinate shifts that interfere with the existing sliding gestures. Consequently, the direct application of finger angle estimation algorithms for recognizing three-dimensional rotational gestures is impractical. SwivelTouch leverages the analysis of finger movement characteristics on the touchscreen captured through original capacitive image sequences, which aims to rapidly and accurately identify these advanced 3D gestures, clearly differentiating them from conventional touch interactions like tapping and sliding, thus enhancing user interaction with touch devices and meanwhile compatible with existing 2D gestures. User study further confirms that the implementation of SwivelTouch significantly enhances the efficiency of text editing on smartphones.
如今,触摸屏已成为移动计算设备(智能手机、平板电脑、智能手表)最普遍的输入设备。然而,与台式机或笔记本电脑相比,触摸屏设备上的快捷键和物理按钮有限,再加上手指肥大的问题,往往导致输入和导航速度更慢、更容易出错,尤其是在处理文本编辑和其他复杂的交互任务时。我们引入了一套基于手指在触摸屏上偏航、俯仰和滚动方向旋转的创新手势,与传统的二维交互方式大相径庭,有望扩展手势库。尽管在手指角度估算方面的研究十分活跃,但之前的工作面临着巨大的挑战,包括显著的估算误差和不稳定的连续输出。用户行为的多变性使将动作隔离到单一旋转轴变得更加复杂,从而导致意外干扰和屏幕坐标偏移,干扰现有的滑动手势。因此,直接应用手指角度估计算法来识别三维旋转手势是不切实际的。SwivelTouch 利用原始电容式图像序列捕获的手指在触摸屏上的运动特征进行分析,旨在快速准确地识别这些高级三维手势,将其与传统的触摸交互(如点击和滑动)明确区分开来,从而增强用户与触摸设备的交互,同时与现有的二维手势兼容。用户研究进一步证实,SwivelTouch 的实施大大提高了智能手机上文本编辑的效率。
{"title":"SwivelTouch: Boosting Touchscreen Input with 3D Finger Rotation Gesture","authors":"Chentao Li, Jinyang Yu, Ke He, Jianjiang Feng, Jie Zhou","doi":"10.1145/3659584","DOIUrl":"https://doi.org/10.1145/3659584","url":null,"abstract":"Today, touchscreens stand as the most prevalent input devices of mobile computing devices (smartphones, tablets, smartwatches). Yet, compared with desktop or laptop computers, the limited shortcut keys and physical buttons on touchscreen devices, coupled with the fat finger problem, often lead to slower and more error-prone input and navigation, especially when dealing with text editing and other complex interaction tasks. We introduce an innovative gesture set based on finger rotations in the yaw, pitch, and roll directions on a touchscreen, diverging significantly from traditional two-dimensional interactions and promising to expand the gesture library. Despite active research in estimation of finger angles, however, the previous work faces substantial challenges, including significant estimation errors and unstable sequential outputs. Variability in user behavior further complicates the isolation of movements to a single rotational axis, leading to accidental disturbances and screen coordinate shifts that interfere with the existing sliding gestures. Consequently, the direct application of finger angle estimation algorithms for recognizing three-dimensional rotational gestures is impractical. SwivelTouch leverages the analysis of finger movement characteristics on the touchscreen captured through original capacitive image sequences, which aims to rapidly and accurately identify these advanced 3D gestures, clearly differentiating them from conventional touch interactions like tapping and sliding, thus enhancing user interaction with touch devices and meanwhile compatible with existing 2D gestures. User study further confirms that the implementation of SwivelTouch significantly enhances the efficiency of text editing on smartphones.","PeriodicalId":20553,"journal":{"name":"Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140983306","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pushing the Limits of Acoustic Spatial Perception via Incident Angle Encoding 通过入射角编码突破声学空间感知的极限
Q1 Computer Science Pub Date : 2024-05-13 DOI: 10.1145/3659583
Yongjian Fu, Yongzhao Zhang, Hao Pan, Yu Lu, Xinyi Li, Lili Chen, Ju Ren, Xiong Li, Xiaosong Zhang, Yaoxue Zhang
With the growing popularity of smart speakers, numerous novel acoustic sensing applications have been proposed for low-frequency human speech and high-frequency inaudible sounds. Spatial information plays a crucial role in these acoustic applications, enabling various location-based services. However, typically commercial microphone arrays face limitations in spatial perception of inaudible sounds due to their sparse array geometries optimized for low-frequency speech. In this paper, we introduce MetaAng, a system designed to augment microphone arrays by enabling wideband spatial perception across both speech signals and inaudible sounds by leveraging the spatial encoding capabilities of acoustic metasurfaces. Our design is grounded in the fact that, while sensitive to high-frequency signals, acoustic metasurfaces are almost non-responsive to low-frequency speech due to significant wavelength discrepancy. This observation allows us to integrate acoustic metasurfaces with sparse array geometry, simultaneously enhancing the spatial perception of high-frequency and low-frequency acoustic signals. To achieve this, we first utilize acoustic metasurfaces and a configuration optimization algorithm to encode the unique features for each incident angle. Then, we propose an unrolling soft thresholding network that employs neural-enhanced priors and compressive sensing for high-accuracy, high-resolution multi-source angle estimation. We implement a prototype, and experimental results demonstrate that MetaAng maintains robustness across various scenarios, facilitating multiple applications, including localization and tracking.
随着智能扬声器的日益普及,针对低频人类语音和高频不可听声音提出了许多新颖的声学传感应用。空间信息在这些声学应用中发挥着至关重要的作用,使各种基于位置的服务成为可能。然而,通常的商用麦克风阵列由于其针对低频语音而优化的稀疏阵列几何结构,在对不可听声音的空间感知方面存在局限性。在本文中,我们介绍了 MetaAng,这是一个旨在增强麦克风阵列的系统,通过利用声学元表面的空间编码能力,实现对语音信号和不可闻声音的宽带空间感知。我们的设计基于以下事实:声学元表面虽然对高频信号很敏感,但由于波长差异很大,对低频语音几乎没有反应。根据这一观察结果,我们可以将声学元表面与稀疏阵列几何相结合,同时增强对高频和低频声学信号的空间感知。为此,我们首先利用声元曲面和配置优化算法来编码每个入射角的独特特征。然后,我们提出了一种开卷软阈值网络,该网络采用神经增强先验和压缩传感技术,用于高精度、高分辨率的多源角度估计。我们实现了一个原型,实验结果表明,MetaAng 在各种情况下都能保持稳健性,为定位和跟踪等多种应用提供了便利。
{"title":"Pushing the Limits of Acoustic Spatial Perception via Incident Angle Encoding","authors":"Yongjian Fu, Yongzhao Zhang, Hao Pan, Yu Lu, Xinyi Li, Lili Chen, Ju Ren, Xiong Li, Xiaosong Zhang, Yaoxue Zhang","doi":"10.1145/3659583","DOIUrl":"https://doi.org/10.1145/3659583","url":null,"abstract":"With the growing popularity of smart speakers, numerous novel acoustic sensing applications have been proposed for low-frequency human speech and high-frequency inaudible sounds. Spatial information plays a crucial role in these acoustic applications, enabling various location-based services. However, typically commercial microphone arrays face limitations in spatial perception of inaudible sounds due to their sparse array geometries optimized for low-frequency speech. In this paper, we introduce MetaAng, a system designed to augment microphone arrays by enabling wideband spatial perception across both speech signals and inaudible sounds by leveraging the spatial encoding capabilities of acoustic metasurfaces. Our design is grounded in the fact that, while sensitive to high-frequency signals, acoustic metasurfaces are almost non-responsive to low-frequency speech due to significant wavelength discrepancy. This observation allows us to integrate acoustic metasurfaces with sparse array geometry, simultaneously enhancing the spatial perception of high-frequency and low-frequency acoustic signals. To achieve this, we first utilize acoustic metasurfaces and a configuration optimization algorithm to encode the unique features for each incident angle. Then, we propose an unrolling soft thresholding network that employs neural-enhanced priors and compressive sensing for high-accuracy, high-resolution multi-source angle estimation. We implement a prototype, and experimental results demonstrate that MetaAng maintains robustness across various scenarios, facilitating multiple applications, including localization and tracking.","PeriodicalId":20553,"journal":{"name":"Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140984062","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Beamforming for Sensing: Hybrid Beamforming based on Transmitter-Receiver Collaboration for Millimeter-Wave Sensing 传感波束成形:基于发射机-接收机协作的混合波束成形用于毫米波传感
Q1 Computer Science Pub Date : 2024-05-13 DOI: 10.1145/3659619
Long Fan, Lei Xie, Wenhui Zhou, Chuyu Wang, Yanling Bu, Sanglu Lu
Previous mmWave sensing solutions assumed good signal quality. Ensuring an unblocked or strengthened LoS path is challenging. Therefore, finding an NLoS path is crucial to enhancing perceived signal quality. This paper proposes Trebsen, a Transmitter-REceiver collaboration-based Beamforming scheme SENsing using commercial mmWave radars. Specifically, we define the hybrid beamforming problem as an optimization challenge involving beamforming angle search based on transmitter-receiver collaboration. We derive a comprehensive expression for parameter optimization by modeling the signal attenuation variations resulting from the propagation path. To comprehensively assess the perception signal quality, we design a novel metric perceived signal-to-interference-plus-noise ratio (PSINR), combining the carrier signal and baseband signal to quantify the fine-grained sensing motion signal quality. Considering the high time cost of traversing or randomly searching methods, we employ a search method based on deep reinforcement learning to quickly explore optimal beamforming angles at both transmitter and receiver. We implement Trebsen and evaluate its performance in a fine-grained sensing application (i.e., heartbeat). Experimental results show that Trebsen significantly enhances heartbeat sensing performance in blocked or misaligned LoS scenes. Comparing non-beamforming, Trebsen demonstrates a reduction of 23.6% in HR error and 27.47% in IBI error. Moreover, comparing random search, Trebsen exhibits a 90% increase in search speed.
以前的毫米波传感解决方案假定信号质量良好。确保无阻塞或增强的 LoS 路径具有挑战性。因此,找到一条 NLoS 路径对于提高感知信号质量至关重要。本文提出的 Trebsen 是一种基于发射机-接收机协作的波束成形方案,利用商用毫米波雷达进行 SENsing。具体来说,我们将混合波束成形问题定义为一个优化挑战,涉及基于发射机-接收机协作的波束成形角度搜索。通过对传播路径造成的信号衰减变化进行建模,我们得出了参数优化的综合表达式。为了全面评估感知信号质量,我们设计了一种新的感知信号干扰加噪声比(PSINR)指标,结合载波信号和基带信号来量化细粒度感知运动信号质量。考虑到遍历或随机搜索方法的时间成本较高,我们采用了一种基于深度强化学习的搜索方法,以快速探索发射器和接收器的最佳波束成形角度。我们实现了 Trebsen,并评估了它在细粒度传感应用(即心跳)中的性能。实验结果表明,Trebsen 显著提高了在阻塞或错位 LoS 场景中的心跳传感性能。与非波束成形相比,Trebsen 将 HR 误差降低了 23.6%,将 IBI 误差降低了 27.47%。此外,与随机搜索相比,Trebsen 的搜索速度提高了 90%。
{"title":"Beamforming for Sensing: Hybrid Beamforming based on Transmitter-Receiver Collaboration for Millimeter-Wave Sensing","authors":"Long Fan, Lei Xie, Wenhui Zhou, Chuyu Wang, Yanling Bu, Sanglu Lu","doi":"10.1145/3659619","DOIUrl":"https://doi.org/10.1145/3659619","url":null,"abstract":"Previous mmWave sensing solutions assumed good signal quality. Ensuring an unblocked or strengthened LoS path is challenging. Therefore, finding an NLoS path is crucial to enhancing perceived signal quality. This paper proposes Trebsen, a Transmitter-REceiver collaboration-based Beamforming scheme SENsing using commercial mmWave radars. Specifically, we define the hybrid beamforming problem as an optimization challenge involving beamforming angle search based on transmitter-receiver collaboration. We derive a comprehensive expression for parameter optimization by modeling the signal attenuation variations resulting from the propagation path. To comprehensively assess the perception signal quality, we design a novel metric perceived signal-to-interference-plus-noise ratio (PSINR), combining the carrier signal and baseband signal to quantify the fine-grained sensing motion signal quality. Considering the high time cost of traversing or randomly searching methods, we employ a search method based on deep reinforcement learning to quickly explore optimal beamforming angles at both transmitter and receiver. We implement Trebsen and evaluate its performance in a fine-grained sensing application (i.e., heartbeat). Experimental results show that Trebsen significantly enhances heartbeat sensing performance in blocked or misaligned LoS scenes. Comparing non-beamforming, Trebsen demonstrates a reduction of 23.6% in HR error and 27.47% in IBI error. Moreover, comparing random search, Trebsen exhibits a 90% increase in search speed.","PeriodicalId":20553,"journal":{"name":"Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140985882","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ClassID: Enabling Student Behavior Attribution from Ambient Classroom Sensing Systems ClassID:通过教室环境感应系统实现学生行为归因
Q1 Computer Science Pub Date : 2024-05-13 DOI: 10.1145/3659586
Prasoon Patidar, Tricia J. Ngoon, John Zimmerman, Amy Ogan, Yuvraj Agarwal
Ambient classroom sensing systems offer a scalable and non-intrusive way to find connections between instructor actions and student behaviors, creating data that can improve teaching and learning. While these systems effectively provide aggregate data, getting reliable individual student-level information is difficult due to occlusion or movements. Individual data can help in understanding equitable student participation, but it requires identifiable data or individual instrumentation. We propose ClassID, a data attribution method for within a class session and across multiple sessions of a course without these constraints. For within-session, our approach assigns unique identifiers to 98% of students with 95% accuracy. It significantly reduces multiple ID assignments compared to the baseline approach (3 vs. 167) based on our testing on data from 15 classroom sessions. For across-session attributions, our approach, combined with student attendance, shows higher precision than the state-of-the-art approach (85% vs. 44%) on three courses. Finally, we present a set of four use cases to demonstrate how individual behavior attribution can enable a rich set of learning analytics, which is not possible with aggregate data alone.
教室环境感知系统提供了一种可扩展的非侵入式方法,可以发现教师行为与学生行为之间的联系,从而创建可改善教学的数据。虽然这些系统能有效提供综合数据,但由于遮挡或移动的原因,很难获得可靠的学生个体信息。个人数据有助于了解学生的公平参与情况,但这需要可识别的数据或个人工具。我们提出的 ClassID 是一种数据归属方法,适用于一节课内和一门课程的多个课时,不受这些限制。在课内,我们的方法为 98% 的学生分配了唯一标识符,准确率达 95%。根据我们对 15 节课堂数据的测试,与基线方法相比,它大大减少了多重 ID 分配(3 对 167)。在跨课程归因方面,我们的方法与学生出勤率相结合,在三门课程上显示出比最先进方法更高的精确度(85% 对 44%)。最后,我们介绍了四个使用案例,以展示个人行为归因如何实现丰富的学习分析,而这是仅靠综合数据无法实现的。
{"title":"ClassID: Enabling Student Behavior Attribution from Ambient Classroom Sensing Systems","authors":"Prasoon Patidar, Tricia J. Ngoon, John Zimmerman, Amy Ogan, Yuvraj Agarwal","doi":"10.1145/3659586","DOIUrl":"https://doi.org/10.1145/3659586","url":null,"abstract":"Ambient classroom sensing systems offer a scalable and non-intrusive way to find connections between instructor actions and student behaviors, creating data that can improve teaching and learning. While these systems effectively provide aggregate data, getting reliable individual student-level information is difficult due to occlusion or movements. Individual data can help in understanding equitable student participation, but it requires identifiable data or individual instrumentation. We propose ClassID, a data attribution method for within a class session and across multiple sessions of a course without these constraints. For within-session, our approach assigns unique identifiers to 98% of students with 95% accuracy. It significantly reduces multiple ID assignments compared to the baseline approach (3 vs. 167) based on our testing on data from 15 classroom sessions. For across-session attributions, our approach, combined with student attendance, shows higher precision than the state-of-the-art approach (85% vs. 44%) on three courses. Finally, we present a set of four use cases to demonstrate how individual behavior attribution can enable a rich set of learning analytics, which is not possible with aggregate data alone.","PeriodicalId":20553,"journal":{"name":"Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140984360","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Predicting Multi-dimensional Surgical Outcomes with Multi-modal Mobile Sensing 利用多模态移动传感技术预测多维手术结果
Q1 Computer Science Pub Date : 2024-05-13 DOI: 10.1145/3659628
Ziqi Xu, Jingwen Zhang, Jacob K Greenberg, Madelyn Frumkin, Saad Javeed, Justin K. Zhang, Braeden Benedict, Kathleen Botterbush, Thomas L. Rodebaugh, Wilson Z. Ray, Chenyang Lu
Pre-operative prediction of post-surgical recovery for patients is vital for clinical decision-making and personalized treatments, especially with lumbar spine surgery, where patients exhibit highly heterogeneous outcomes. Existing predictive tools mainly rely on traditional Patient-Reported Outcome Measures (PROMs), which fail to capture the long-term dynamics of patient conditions before the surgery. Moreover, existing studies focus on predicting a single surgical outcome. However, recovery from spine surgery is multi-dimensional, including multiple distinctive but interrelated outcomes, such as pain interference, physical function, and quality of recovery. In recent years, the emergence of smartphones and wearable devices has presented new opportunities to capture longitudinal and dynamic information regarding patients' conditions outside the hospital. This paper proposes a novel machine learning approach, Multi-Modal Multi-Task Learning (M3TL), using smartphones and wristbands to predict multiple surgical outcomes after lumbar spine surgeries. We formulate the prediction of pain interference, physical function, and quality of recovery as a multi-task learning (MTL) problem. We leverage multi-modal data to capture the static and dynamic characteristics of patients, including (1) traditional features from PROMs and Electronic Health Records (EHR), (2) Ecological Momentary Assessment (EMA) collected from smartphones, and (3) sensing data from wristbands. Moreover, we introduce new features derived from the correlation of EMA and wearable features measured within the same time frame, effectively enhancing predictive performance by capturing the interdependencies between the two data modalities. Our model interpretation uncovers the complementary nature of the different data modalities and their distinctive contributions toward multiple surgical outcomes. Furthermore, through individualized decision analysis, our model identifies personal high risk factors to aid clinical decision making and approach personalized treatments. In a clinical study involving 122 patients undergoing lumbar spine surgery, our M3TL model outperforms a diverse set of baseline methods in predictive performance, demonstrating the value of integrating multi-modal data and learning from multiple surgical outcomes. This work contributes to advancing personalized peri-operative care with accurate pre-operative predictions of multi-dimensional outcomes.
术前预测患者的术后恢复情况对于临床决策和个性化治疗至关重要,尤其是腰椎手术,因为患者的术后恢复情况千差万别。现有的预测工具主要依赖于传统的患者报告结果指标(PROMs),而这些指标无法捕捉到患者术前病情的长期动态变化。此外,现有研究侧重于预测单一的手术结果。然而,脊柱手术后的恢复是多维度的,包括疼痛干扰、身体功能和恢复质量等多个不同但相互关联的结果。近年来,智能手机和可穿戴设备的出现为在医院外获取有关患者病情的纵向动态信息提供了新的机遇。本文提出了一种新颖的机器学习方法--多模态多任务学习(M3TL),利用智能手机和腕带来预测腰椎手术后的多种手术结果。我们将疼痛干扰、身体功能和恢复质量的预测制定为一个多任务学习(MTL)问题。我们利用多模态数据来捕捉患者的静态和动态特征,包括:(1)来自PROM和电子健康记录(EHR)的传统特征;(2)从智能手机收集的生态瞬间评估(EMA);以及(3)来自腕带的传感数据。此外,我们还引入了在同一时间段内测量的 EMA 和可穿戴设备特征的相关性所产生的新特征,通过捕捉两种数据模式之间的相互依存关系,有效提高了预测性能。我们的模型解释揭示了不同数据模式的互补性及其对多种手术结果的独特贡献。此外,通过个性化决策分析,我们的模型还能识别个人高风险因素,以帮助临床决策和个性化治疗。在一项涉及 122 名腰椎手术患者的临床研究中,我们的 M3TL 模型在预测性能方面优于各种基线方法,证明了整合多模态数据并从多种手术结果中学习的价值。这项工作有助于通过对多维结果进行准确的术前预测来推进个性化围手术期护理。
{"title":"Predicting Multi-dimensional Surgical Outcomes with Multi-modal Mobile Sensing","authors":"Ziqi Xu, Jingwen Zhang, Jacob K Greenberg, Madelyn Frumkin, Saad Javeed, Justin K. Zhang, Braeden Benedict, Kathleen Botterbush, Thomas L. Rodebaugh, Wilson Z. Ray, Chenyang Lu","doi":"10.1145/3659628","DOIUrl":"https://doi.org/10.1145/3659628","url":null,"abstract":"Pre-operative prediction of post-surgical recovery for patients is vital for clinical decision-making and personalized treatments, especially with lumbar spine surgery, where patients exhibit highly heterogeneous outcomes. Existing predictive tools mainly rely on traditional Patient-Reported Outcome Measures (PROMs), which fail to capture the long-term dynamics of patient conditions before the surgery. Moreover, existing studies focus on predicting a single surgical outcome. However, recovery from spine surgery is multi-dimensional, including multiple distinctive but interrelated outcomes, such as pain interference, physical function, and quality of recovery. In recent years, the emergence of smartphones and wearable devices has presented new opportunities to capture longitudinal and dynamic information regarding patients' conditions outside the hospital. This paper proposes a novel machine learning approach, Multi-Modal Multi-Task Learning (M3TL), using smartphones and wristbands to predict multiple surgical outcomes after lumbar spine surgeries. We formulate the prediction of pain interference, physical function, and quality of recovery as a multi-task learning (MTL) problem. We leverage multi-modal data to capture the static and dynamic characteristics of patients, including (1) traditional features from PROMs and Electronic Health Records (EHR), (2) Ecological Momentary Assessment (EMA) collected from smartphones, and (3) sensing data from wristbands. Moreover, we introduce new features derived from the correlation of EMA and wearable features measured within the same time frame, effectively enhancing predictive performance by capturing the interdependencies between the two data modalities. Our model interpretation uncovers the complementary nature of the different data modalities and their distinctive contributions toward multiple surgical outcomes. Furthermore, through individualized decision analysis, our model identifies personal high risk factors to aid clinical decision making and approach personalized treatments. In a clinical study involving 122 patients undergoing lumbar spine surgery, our M3TL model outperforms a diverse set of baseline methods in predictive performance, demonstrating the value of integrating multi-modal data and learning from multiple surgical outcomes. This work contributes to advancing personalized peri-operative care with accurate pre-operative predictions of multi-dimensional outcomes.","PeriodicalId":20553,"journal":{"name":"Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140984997","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1