首页 > 最新文献

arXiv - CS - Human-Computer Interaction最新文献

英文 中文
Charting EDA: Characterizing Interactive Visualization Use in Computational Notebooks with a Mixed-Methods Formalism 绘制 EDA 图表:用混合方法表征计算笔记本中交互式可视化的使用情况
Pub Date : 2024-09-16 DOI: arxiv-2409.10450
Dylan Wootton, Amy Rae Fox, Evan Peck, Arvind Satyanarayan
Interactive visualizations are powerful tools for Exploratory Data Analysis(EDA), but how do they affect the observations analysts make about their data?We conducted a qualitative experiment with 13 professional data scientistsanalyzing two datasets with Jupyter notebooks, collecting a rich dataset ofinteraction traces and think-aloud utterances. By qualitatively codingparticipant utterances, we introduce a formalism that describes EDA as asequence of analysis states, where each state is comprised of either arepresentation an analyst constructs (e.g., the output of a data frame, aninteractive visualization, etc.) or an observation the analyst makes (e.g.,about missing data, the relationship between variables, etc.). By applying ourformalism to our dataset, we identify that interactive visualizations, onaverage, lead to earlier and more complex insights about relationships betweendataset attributes compared to static visualizations. Moreover, by calculatingmetrics such as revisit count and representational diversity, we uncover thatsome representations serve more as "planning aids" during EDA rather than toolsstrictly for hypothesis-answering. We show how these measures help identifyother patterns of analysis behavior, such as the "80-20 rule", where a smallsubset of representations drove the majority of observations. Based on thesefindings, we offer design guidelines for interactive exploratory analysistooling and reflect on future directions for studying the role thatvisualizations play in EDA.
交互式可视化是探索性数据分析(EDA)的强大工具,但交互式可视化如何影响分析师对数据的观察?通过对参与者的话语进行定性编码,我们引入了一种形式主义,将 EDA 描述为一系列分析状态,其中每个状态都由分析师构建的表述(如数据框架的输出、交互式可视化等)或分析师的观察(如关于缺失数据、变量之间的关系等)组成。通过将形式主义应用于数据集,我们发现,与静态可视化相比,交互式可视化平均能更早更复杂地洞察数据集属性之间的关系。此外,通过计算重访次数和表征多样性等指标,我们发现有些表征更像是 EDA 过程中的 "规划辅助工具",而非严格意义上的假设解答工具。我们展示了这些指标如何帮助识别分析行为的其他模式,例如 "80-20 规则",即一小部分表征驱动了大部分观察。基于这些发现,我们为交互式探索分析工具提供了设计指南,并思考了研究可视化在 EDA 中的作用的未来方向。
{"title":"Charting EDA: Characterizing Interactive Visualization Use in Computational Notebooks with a Mixed-Methods Formalism","authors":"Dylan Wootton, Amy Rae Fox, Evan Peck, Arvind Satyanarayan","doi":"arxiv-2409.10450","DOIUrl":"https://doi.org/arxiv-2409.10450","url":null,"abstract":"Interactive visualizations are powerful tools for Exploratory Data Analysis\u0000(EDA), but how do they affect the observations analysts make about their data?\u0000We conducted a qualitative experiment with 13 professional data scientists\u0000analyzing two datasets with Jupyter notebooks, collecting a rich dataset of\u0000interaction traces and think-aloud utterances. By qualitatively coding\u0000participant utterances, we introduce a formalism that describes EDA as a\u0000sequence of analysis states, where each state is comprised of either a\u0000representation an analyst constructs (e.g., the output of a data frame, an\u0000interactive visualization, etc.) or an observation the analyst makes (e.g.,\u0000about missing data, the relationship between variables, etc.). By applying our\u0000formalism to our dataset, we identify that interactive visualizations, on\u0000average, lead to earlier and more complex insights about relationships between\u0000dataset attributes compared to static visualizations. Moreover, by calculating\u0000metrics such as revisit count and representational diversity, we uncover that\u0000some representations serve more as \"planning aids\" during EDA rather than tools\u0000strictly for hypothesis-answering. We show how these measures help identify\u0000other patterns of analysis behavior, such as the \"80-20 rule\", where a small\u0000subset of representations drove the majority of observations. Based on these\u0000findings, we offer design guidelines for interactive exploratory analysis\u0000tooling and reflect on future directions for studying the role that\u0000visualizations play in EDA.","PeriodicalId":501541,"journal":{"name":"arXiv - CS - Human-Computer Interaction","volume":"210 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142269799","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Protocol for identifying shared articulatory features of gestures and LSF: application to epistemic gesture 识别手势和LSF共同发音特征的协议:应用于认识手势
Pub Date : 2024-09-16 DOI: arxiv-2409.10079
Fanny CatteauSFL, Claudia S BianchiniFoReLLIS
This article focuses on the articulatory characteristics of epistemicgestures (i.e., gestures used to express certainty or uncertainty) in co-speechgestures (CSG) in French and in French Sign Language (LSF). It presents a newmethodology for analysis, which relies on the complementary use of manualannotation (using Typannot) and semi-automatic annotation (using AlphaPose) tohighlight the kinesiological characteristics of these epistemic gestures. Thepresented methodology allows to analyze the flexion/extension movements of thehead in epistemic contexts. The results of this analysis show that in CSG andLSF: (1) head nods passing through the neutral position (i.e., head straightwith no flexion/extension) and high movement speed are markers of certainty;and (2) holding the head position away from the neutral position and lowmovement speed indicate uncertainty. This study is conducted within theframework of the ANR LexiKHuM project, which develops kinesthetic communicationsolutions for human-machine interaction.
本文重点研究了法语和法语手语(LSF)共同言语手势(CSG)中认识性手势(即用于表达确定性或不确定性的手势)的发音特征。它提出了一种新的分析方法,该方法依赖于人工标注(使用 Typannot)和半自动标注(使用 AlphaPose)的互补使用,以突出这些认识论手势的运动学特征。本文介绍的方法可用于分析认识语境中头部的屈伸运动。分析结果表明,在 CSG 和LSF 中:(1) 通过中立位置(即头伸直无屈伸)的点头和高运动速度是确定性的标志;(2) 头部位置偏离中立位置和低运动速度表示不确定性。这项研究是在 ANR LexiKHuM 项目框架内进行的,该项目旨在开发用于人机交互的动觉通信解决方案。
{"title":"Protocol for identifying shared articulatory features of gestures and LSF: application to epistemic gesture","authors":"Fanny CatteauSFL, Claudia S BianchiniFoReLLIS","doi":"arxiv-2409.10079","DOIUrl":"https://doi.org/arxiv-2409.10079","url":null,"abstract":"This article focuses on the articulatory characteristics of epistemic\u0000gestures (i.e., gestures used to express certainty or uncertainty) in co-speech\u0000gestures (CSG) in French and in French Sign Language (LSF). It presents a new\u0000methodology for analysis, which relies on the complementary use of manual\u0000annotation (using Typannot) and semi-automatic annotation (using AlphaPose) to\u0000highlight the kinesiological characteristics of these epistemic gestures. The\u0000presented methodology allows to analyze the flexion/extension movements of the\u0000head in epistemic contexts. The results of this analysis show that in CSG and\u0000LSF: (1) head nods passing through the neutral position (i.e., head straight\u0000with no flexion/extension) and high movement speed are markers of certainty;\u0000and (2) holding the head position away from the neutral position and low\u0000movement speed indicate uncertainty. This study is conducted within the\u0000framework of the ANR LexiKHuM project, which develops kinesthetic communication\u0000solutions for human-machine interaction.","PeriodicalId":501541,"journal":{"name":"arXiv - CS - Human-Computer Interaction","volume":"16 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142252475","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Co-Designing Dynamic Mixed Reality Drill Positioning Widgets: A Collaborative Approach with Dentists in a Realistic Setup 共同设计动态混合现实牙钻定位小工具:在现实环境中与牙医合作的方法
Pub Date : 2024-09-16 DOI: arxiv-2409.10258
Mine Dastan, Michele Fiorentino, Elias D. Walter, Christian Diegritz, Antonio E. Uva, Ulrich Eck, Nassir Navab
Mixed Reality (MR) is proven in the literature to support precise spatialdental drill positioning by superimposing 3D widgets. Despite this, the relatedknowledge about widget's visual design and interactive user feedback is stilllimited. Therefore, this study is contributed to by co-designed MR drill toolpositioning widgets with two expert dentists and three MR experts. The resultsof co-design are two static widgets (SWs): a simple entry point, a target axis,and two dynamic widgets (DWs), variants of dynamic error visualization with andwithout a target axis (DWTA and DWEP). We evaluated the co-designed widgets ina virtual reality simulation supported by a realistic setup with a trackedphantom patient, a virtual magnifying loupe, and a dentist's foot pedal. Theuser study involved 35 dentists with various backgrounds and years ofexperience. The findings demonstrated significant results; DWs outperform SWsin positional and rotational precision, especially with younger generations andsubjects with gaming experiences. The user preference remains for DWs (19)instead of SWs (16). However, findings indicated that the precision positivelycorrelates with the time trade-off. The post-experience questionnaire(NASA-TLX) showed that DWs increase mental and physical demand, effort, andfrustration more than SWs. Comparisons between DWEP and DWTA show that the DW'scomplexity level influences time, physical and mental demands. The DWs areextensible to diverse medical and industrial scenarios that demand precision.
文献证明,混合现实(MR)可以通过叠加三维小部件来支持精确的牙钻空间定位。尽管如此,关于小部件的视觉设计和交互式用户反馈的相关知识仍然有限。因此,本研究与两位牙科专家和三位磁共振专家共同设计了磁共振钻头工具定位小部件。共同设计的成果包括两个静态小部件(SW):一个简单的入口点、一个目标轴,以及两个动态小部件(DW):有目标轴和无目标轴的动态误差可视化变体(DWTA 和 DWEP)。我们在虚拟现实模拟中对共同设计的小部件进行了评估,该虚拟现实模拟由一个带有跟踪象鼻病人、虚拟放大镜和牙医脚踏板的逼真装置提供支持。这项用户研究涉及 35 名牙医,他们的背景和经验各不相同。研究结果表明,DW 在位置和旋转精度方面优于 SW,尤其是在年轻一代和有游戏经验的受试者中。用户仍然倾向于使用 DW(19)而不是 SW(16)。不过,研究结果表明,精度与时间权衡呈正相关。体验后问卷(NASA-TLX)显示,DWs 比 SWs 更能增加身心需求、努力和挫败感。DWEP 和 DWTA 之间的比较表明,DW 的复杂程度影响着时间、体力和脑力需求。DW 可扩展到要求精确的各种医疗和工业场景。
{"title":"Co-Designing Dynamic Mixed Reality Drill Positioning Widgets: A Collaborative Approach with Dentists in a Realistic Setup","authors":"Mine Dastan, Michele Fiorentino, Elias D. Walter, Christian Diegritz, Antonio E. Uva, Ulrich Eck, Nassir Navab","doi":"arxiv-2409.10258","DOIUrl":"https://doi.org/arxiv-2409.10258","url":null,"abstract":"Mixed Reality (MR) is proven in the literature to support precise spatial\u0000dental drill positioning by superimposing 3D widgets. Despite this, the related\u0000knowledge about widget's visual design and interactive user feedback is still\u0000limited. Therefore, this study is contributed to by co-designed MR drill tool\u0000positioning widgets with two expert dentists and three MR experts. The results\u0000of co-design are two static widgets (SWs): a simple entry point, a target axis,\u0000and two dynamic widgets (DWs), variants of dynamic error visualization with and\u0000without a target axis (DWTA and DWEP). We evaluated the co-designed widgets in\u0000a virtual reality simulation supported by a realistic setup with a tracked\u0000phantom patient, a virtual magnifying loupe, and a dentist's foot pedal. The\u0000user study involved 35 dentists with various backgrounds and years of\u0000experience. The findings demonstrated significant results; DWs outperform SWs\u0000in positional and rotational precision, especially with younger generations and\u0000subjects with gaming experiences. The user preference remains for DWs (19)\u0000instead of SWs (16). However, findings indicated that the precision positively\u0000correlates with the time trade-off. The post-experience questionnaire\u0000(NASA-TLX) showed that DWs increase mental and physical demand, effort, and\u0000frustration more than SWs. Comparisons between DWEP and DWTA show that the DW's\u0000complexity level influences time, physical and mental demands. The DWs are\u0000extensible to diverse medical and industrial scenarios that demand precision.","PeriodicalId":501541,"journal":{"name":"arXiv - CS - Human-Computer Interaction","volume":"192 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142252472","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Model-in-the-Loop (MILO): Accelerating Multimodal AI Data Annotation with LLMs 环中模型(MILO):利用 LLM 加速多模态人工智能数据注释
Pub Date : 2024-09-16 DOI: arxiv-2409.10702
Yifan Wang, David Stevens, Pranay Shah, Wenwen Jiang, Miao Liu, Xu Chen, Robert Kuo, Na Li, Boying Gong, Daniel Lee, Jiabo Hu, Ning Zhang, Bob Kamma
The growing demand for AI training data has transformed data annotation intoa global industry, but traditional approaches relying on human annotators areoften time-consuming, labor-intensive, and prone to inconsistent quality. Wepropose the Model-in-the-Loop (MILO) framework, which integrates AI/ML modelsinto the annotation process. Our research introduces a collaborative paradigmthat leverages the strengths of both professional human annotators and largelanguage models (LLMs). By employing LLMs as pre-annotation and real-timeassistants, and judges on annotator responses, MILO enables effectiveinteraction patterns between human annotators and LLMs. Three empirical studieson multimodal data annotation demonstrate MILO's efficacy in reducing handlingtime, improving data quality, and enhancing annotator experiences. We alsointroduce quality rubrics for flexible evaluation and fine-grained feedback onopen-ended annotations. The MILO framework has implications for acceleratingAI/ML development, reducing reliance on human annotation alone, and promotingbetter alignment between human and machine values.
对人工智能训练数据日益增长的需求已将数据标注转变为一项全球性产业,但依赖人工标注员的传统方法往往耗时、耗力,而且容易出现质量不一致的问题。我们提出了 "环中模型"(MILO)框架,将人工智能/ML 模型集成到注释过程中。我们的研究引入了一种协作范式,充分利用了专业人工标注员和大型语言模型(LLM)的优势。通过使用 LLM 作为预注释和实时助手,以及对注释者的反应进行评判,MILO 实现了人类注释者和 LLM 之间的有效交互模式。三项关于多模态数据标注的实证研究证明了 MILO 在减少处理时间、提高数据质量和增强标注者体验方面的功效。我们还引入了质量评分标准,以便对开放式注释进行灵活评估和精细反馈。MILO 框架对加速人工智能/ML 开发、减少对人工注释的依赖以及促进人与机器价值之间更好的协调都具有重要意义。
{"title":"Model-in-the-Loop (MILO): Accelerating Multimodal AI Data Annotation with LLMs","authors":"Yifan Wang, David Stevens, Pranay Shah, Wenwen Jiang, Miao Liu, Xu Chen, Robert Kuo, Na Li, Boying Gong, Daniel Lee, Jiabo Hu, Ning Zhang, Bob Kamma","doi":"arxiv-2409.10702","DOIUrl":"https://doi.org/arxiv-2409.10702","url":null,"abstract":"The growing demand for AI training data has transformed data annotation into\u0000a global industry, but traditional approaches relying on human annotators are\u0000often time-consuming, labor-intensive, and prone to inconsistent quality. We\u0000propose the Model-in-the-Loop (MILO) framework, which integrates AI/ML models\u0000into the annotation process. Our research introduces a collaborative paradigm\u0000that leverages the strengths of both professional human annotators and large\u0000language models (LLMs). By employing LLMs as pre-annotation and real-time\u0000assistants, and judges on annotator responses, MILO enables effective\u0000interaction patterns between human annotators and LLMs. Three empirical studies\u0000on multimodal data annotation demonstrate MILO's efficacy in reducing handling\u0000time, improving data quality, and enhancing annotator experiences. We also\u0000introduce quality rubrics for flexible evaluation and fine-grained feedback on\u0000open-ended annotations. The MILO framework has implications for accelerating\u0000AI/ML development, reducing reliance on human annotation alone, and promoting\u0000better alignment between human and machine values.","PeriodicalId":501541,"journal":{"name":"arXiv - CS - Human-Computer Interaction","volume":"54 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142252431","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Algorithmic Behaviors Across Regions: A Geolocation Audit of YouTube Search for COVID-19 Misinformation between the United States and South Africa 跨地区的算法行为:对美国和南非之间 YouTube 搜索 COVID-19 错误信息的地理定位审计
Pub Date : 2024-09-16 DOI: arxiv-2409.10168
Hayoung Jung, Prerna Juneja, Tanushree Mitra
Despite being an integral tool for finding health-related information online,YouTube has faced criticism for disseminating COVID-19 misinformation globallyto its users. Yet, prior audit studies have predominantly investigated YouTubewithin the Global North contexts, often overlooking the Global South. Toaddress this gap, we conducted a comprehensive 10-day geolocation-based auditon YouTube to compare the prevalence of COVID-19 misinformation in searchresults between the United States (US) and South Africa (SA), the countriesheavily affected by the pandemic in the Global North and the Global South,respectively. For each country, we selected 3 geolocations and placedsock-puppets, or bots emulating "real" users, that collected search results for48 search queries sorted by 4 search filters for 10 days, yielding a dataset of915K results. We found that 31.55% of the top-10 search results containedCOVID-19 misinformation. Among the top-10 search results, bots in SA facedsignificantly more misinformative search results than their US counterparts.Overall, our study highlights the contrasting algorithmic behaviors of YouTubesearch between two countries, underscoring the need for the platform toregulate algorithmic behavior consistently across different regions of theGlobe.
尽管YouTube是人们在网上查找健康相关信息的一个不可或缺的工具,但它却因在全球范围内向用户传播COVID-19错误信息而饱受批评。然而,以往的审计研究主要调查的是全球北方背景下的 YouTube,往往忽略了全球南方的情况。为了填补这一空白,我们在 YouTube 上进行了为期 10 天的基于地理位置的全面审核,以比较美国和南非搜索结果中 COVID-19 错误信息的流行程度,这两个国家分别是全球北方和全球南方受该流行病影响最严重的国家。我们为每个国家选择了 3 个地理位置,并放置了模拟 "真实 "用户的 "袜子傀儡 "或机器人,这些机器人收集了 48 个搜索查询的搜索结果,并按 4 个搜索过滤器进行了为期 10 天的排序,从而生成了一个包含 91.5 万个搜索结果的数据集。我们发现,在排名前十的搜索结果中,有 31.55% 含有《COVID-19》的错误信息。总体而言,我们的研究强调了两个国家之间优酷搜索算法行为的反差,强调了平台在全球不同地区对算法行为进行一致监管的必要性。
{"title":"Algorithmic Behaviors Across Regions: A Geolocation Audit of YouTube Search for COVID-19 Misinformation between the United States and South Africa","authors":"Hayoung Jung, Prerna Juneja, Tanushree Mitra","doi":"arxiv-2409.10168","DOIUrl":"https://doi.org/arxiv-2409.10168","url":null,"abstract":"Despite being an integral tool for finding health-related information online,\u0000YouTube has faced criticism for disseminating COVID-19 misinformation globally\u0000to its users. Yet, prior audit studies have predominantly investigated YouTube\u0000within the Global North contexts, often overlooking the Global South. To\u0000address this gap, we conducted a comprehensive 10-day geolocation-based audit\u0000on YouTube to compare the prevalence of COVID-19 misinformation in search\u0000results between the United States (US) and South Africa (SA), the countries\u0000heavily affected by the pandemic in the Global North and the Global South,\u0000respectively. For each country, we selected 3 geolocations and placed\u0000sock-puppets, or bots emulating \"real\" users, that collected search results for\u000048 search queries sorted by 4 search filters for 10 days, yielding a dataset of\u0000915K results. We found that 31.55% of the top-10 search results contained\u0000COVID-19 misinformation. Among the top-10 search results, bots in SA faced\u0000significantly more misinformative search results than their US counterparts.\u0000Overall, our study highlights the contrasting algorithmic behaviors of YouTube\u0000search between two countries, underscoring the need for the platform to\u0000regulate algorithmic behavior consistently across different regions of the\u0000Globe.","PeriodicalId":501541,"journal":{"name":"arXiv - CS - Human-Computer Interaction","volume":"42 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142252527","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On the Effect of Robot Errors on Human Teaching Dynamics 论机器人错误对人类教学动力的影响
Pub Date : 2024-09-15 DOI: arxiv-2409.09827
Jindan Huang, Isaac Sheidlower, Reuben M. Aronson, Elaine Schaertl Short
Human-in-the-loop learning is gaining popularity, particularly in the fieldof robotics, because it leverages human knowledge about real-world tasks tofacilitate agent learning. When people instruct robots, they naturally adapttheir teaching behavior in response to changes in robot performance. Whilecurrent research predominantly focuses on integrating human teaching dynamicsfrom an algorithmic perspective, understanding these dynamics from ahuman-centered standpoint is an under-explored, yet fundamental problem.Addressing this issue will enhance both robot learning and user experience.Therefore, this paper explores one potential factor contributing to the dynamicnature of human teaching: robot errors. We conducted a user study toinvestigate how the presence and severity of robot errors affect threedimensions of human teaching dynamics: feedback granularity, feedback richness,and teaching time, in both forced-choice and open-ended teaching contexts. Theresults show that people tend to spend more time teaching robots with errors,provide more detailed feedback over specific segments of a robot's trajectory,and that robot error can influence a teacher's choice of feedback modality. Ourfindings offer valuable insights for designing effective interfaces forinteractive learning and optimizing algorithms to better understand humanintentions.
人在回路中学习越来越受欢迎,尤其是在机器人领域,因为它利用人类对真实世界任务的了解来促进机器人的学习。人在指导机器人时,自然会根据机器人性能的变化调整自己的教学行为。虽然目前的研究主要侧重于从算法角度整合人类的教学动态,但从以人为本的角度理解这些动态是一个尚未得到充分探索的基本问题。因此,本文探讨了导致人类教学动态的一个潜在因素:机器人错误。我们进行了一项用户研究,以调查机器人错误的存在和严重程度如何影响人类教学动态的三个维度:在强制选择和开放式教学情境中的反馈粒度、反馈丰富度和教学时间。结果表明,人类倾向于花更多时间教授有错误的机器人,对机器人轨迹的特定部分提供更详细的反馈,而且机器人错误会影响教师对反馈方式的选择。我们的发现为设计有效的交互式学习界面和优化算法以更好地理解人类意图提供了宝贵的见解。
{"title":"On the Effect of Robot Errors on Human Teaching Dynamics","authors":"Jindan Huang, Isaac Sheidlower, Reuben M. Aronson, Elaine Schaertl Short","doi":"arxiv-2409.09827","DOIUrl":"https://doi.org/arxiv-2409.09827","url":null,"abstract":"Human-in-the-loop learning is gaining popularity, particularly in the field\u0000of robotics, because it leverages human knowledge about real-world tasks to\u0000facilitate agent learning. When people instruct robots, they naturally adapt\u0000their teaching behavior in response to changes in robot performance. While\u0000current research predominantly focuses on integrating human teaching dynamics\u0000from an algorithmic perspective, understanding these dynamics from a\u0000human-centered standpoint is an under-explored, yet fundamental problem.\u0000Addressing this issue will enhance both robot learning and user experience.\u0000Therefore, this paper explores one potential factor contributing to the dynamic\u0000nature of human teaching: robot errors. We conducted a user study to\u0000investigate how the presence and severity of robot errors affect three\u0000dimensions of human teaching dynamics: feedback granularity, feedback richness,\u0000and teaching time, in both forced-choice and open-ended teaching contexts. The\u0000results show that people tend to spend more time teaching robots with errors,\u0000provide more detailed feedback over specific segments of a robot's trajectory,\u0000and that robot error can influence a teacher's choice of feedback modality. Our\u0000findings offer valuable insights for designing effective interfaces for\u0000interactive learning and optimizing algorithms to better understand human\u0000intentions.","PeriodicalId":501541,"journal":{"name":"arXiv - CS - Human-Computer Interaction","volume":"110 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142268607","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ExploreSelf: Fostering User-driven Exploration and Reflection on Personal Challenges with Adaptive Guidance by Large Language Models 探索自我:通过大型语言模型的自适应指导,促进用户对个人挑战进行探索和反思
Pub Date : 2024-09-15 DOI: arxiv-2409.09662
Inhwa Song, SoHyun Park, Sachin R. Pendse, Jessica Lee Schleider, Munmun De Choudhury, Young-Ho Kim
Expressing stressful experiences in words is proven to improve mental andphysical health, but individuals often disengage with writing interventions asthey struggle to organize their thoughts and emotions. Reflective prompts havebeen used to provide direction, and large language models (LLMs) havedemonstrated the potential to provide tailored guidance. Current systems oftenlimit users' flexibility to direct their reflections. We thus presentExploreSelf, an LLM-driven application designed to empower users to controltheir reflective journey. ExploreSelf allows users to receive adaptive supportthrough dynamically generated questions. Through an exploratory study with 19participants, we examine how participants explore and reflect on personalchallenges using ExploreSelf. Our findings demonstrate that participants valuedthe balance between guided support and freedom to control their reflectivejourney, leading to deeper engagement and insight. Building on our findings, wediscuss implications for designing LLM-driven tools that promote userempowerment through effective reflective practices.
事实证明,用语言表达压力体验可以改善身心健康,但个人往往会因为难以组织自己的思想和情绪而放弃写作干预。反思性提示已被用于提供指导,而大型语言模型(LLM)已证明了提供定制指导的潜力。目前的系统往往限制了用户引导反思的灵活性。因此,我们提出了探索自我(ExploreSelf),这是一个由 LLM 驱动的应用程序,旨在让用户有能力控制自己的反思之旅。ExploreSelf允许用户通过动态生成的问题获得自适应支持。通过对 19 名参与者的探索性研究,我们考察了参与者如何使用 ExploreSelf 探索和反思个人挑战。我们的研究结果表明,参与者非常重视在指导支持和自由控制反思旅程之间取得平衡,从而获得更深的参与感和洞察力。在研究结果的基础上,我们讨论了设计以 LLM 为驱动的工具的意义,这些工具通过有效的反思实践促进了用户赋权。
{"title":"ExploreSelf: Fostering User-driven Exploration and Reflection on Personal Challenges with Adaptive Guidance by Large Language Models","authors":"Inhwa Song, SoHyun Park, Sachin R. Pendse, Jessica Lee Schleider, Munmun De Choudhury, Young-Ho Kim","doi":"arxiv-2409.09662","DOIUrl":"https://doi.org/arxiv-2409.09662","url":null,"abstract":"Expressing stressful experiences in words is proven to improve mental and\u0000physical health, but individuals often disengage with writing interventions as\u0000they struggle to organize their thoughts and emotions. Reflective prompts have\u0000been used to provide direction, and large language models (LLMs) have\u0000demonstrated the potential to provide tailored guidance. Current systems often\u0000limit users' flexibility to direct their reflections. We thus present\u0000ExploreSelf, an LLM-driven application designed to empower users to control\u0000their reflective journey. ExploreSelf allows users to receive adaptive support\u0000through dynamically generated questions. Through an exploratory study with 19\u0000participants, we examine how participants explore and reflect on personal\u0000challenges using ExploreSelf. Our findings demonstrate that participants valued\u0000the balance between guided support and freedom to control their reflective\u0000journey, leading to deeper engagement and insight. Building on our findings, we\u0000discuss implications for designing LLM-driven tools that promote user\u0000empowerment through effective reflective practices.","PeriodicalId":501541,"journal":{"name":"arXiv - CS - Human-Computer Interaction","volume":"46 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142268567","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MindScape Study: Integrating LLM and Behavioral Sensing for Personalized AI-Driven Journaling Experiences MindScape 研究:整合 LLM 和行为传感,打造个性化人工智能驱动的日志体验
Pub Date : 2024-09-15 DOI: arxiv-2409.09570
Subigya Nepal, Arvind Pillai, William Campbell, Talie Massachi, Michael V. Heinz, Ashmita Kunwar, Eunsol Soul Choi, Orson Xu, Joanna Kuc, Jeremy Huckins, Jason Holden, Sarah M. Preum, Colin Depp, Nicholas Jacobson, Mary Czerwinski, Eric Granholm, Andrew T. Campbell
Mental health concerns are prevalent among college students, highlighting theneed for effective interventions that promote self-awareness and holisticwell-being. MindScape pioneers a novel approach to AI-powered journaling byintegrating passively collected behavioral patterns such as conversationalengagement, sleep, and location with Large Language Models (LLMs). Thisintegration creates a highly personalized and context-aware journalingexperience, enhancing self-awareness and well-being by embedding behavioralintelligence into AI. We present an 8-week exploratory study with 20 collegestudents, demonstrating the MindScape app's efficacy in enhancing positiveaffect (7%), reducing negative affect (11%), loneliness (6%), and anxiety anddepression, with a significant week-over-week decrease in PHQ-4 scores (-0.25coefficient), alongside improvements in mindfulness (7%) and self-reflection(6%). The study highlights the advantages of contextual AI journaling, withparticipants particularly appreciating the tailored prompts and insightsprovided by the MindScape app. Our analysis also includes a comparison ofresponses to AI-driven contextual versus generic prompts, participant feedbackinsights, and proposed strategies for leveraging contextual AI journaling toimprove well-being on college campuses. By showcasing the potential ofcontextual AI journaling to support mental health, we provide a foundation forfurther investigation into the effects of contextual AI journaling on mentalhealth and well-being.
心理健康问题在大学生中十分普遍,这凸显了对促进自我意识和整体健康的有效干预的需求。MindScape 通过将被动收集的行为模式(如对话参与、睡眠和位置)与大型语言模型(LLMs)相结合,开创了一种人工智能驱动的日记新方法。这种整合创造了一种高度个性化和情境感知的日志体验,通过将行为智能嵌入人工智能来提高自我意识和幸福感。我们对 20 名大学生进行了为期 8 周的探索性研究,结果表明 MindScape 应用程序在增强积极影响(7%)、减少消极影响(11%)、孤独感(6%)、焦虑和抑郁方面具有功效,PHQ-4 分数一周比一周显著下降(系数为-0.25),正念(7%)和自我反省(6%)也有所改善。这项研究强调了情境式人工智能日志的优势,参与者尤其欣赏 MindScape 应用程序提供的定制提示和见解。我们的分析还包括对人工智能驱动的情境提示与一般提示的反应、参与者的反馈意见以及利用情境人工智能日志改善大学校园幸福感的建议策略进行比较。通过展示情境式人工智能日志支持心理健康的潜力,我们为进一步研究情境式人工智能日志对心理健康和幸福感的影响奠定了基础。
{"title":"MindScape Study: Integrating LLM and Behavioral Sensing for Personalized AI-Driven Journaling Experiences","authors":"Subigya Nepal, Arvind Pillai, William Campbell, Talie Massachi, Michael V. Heinz, Ashmita Kunwar, Eunsol Soul Choi, Orson Xu, Joanna Kuc, Jeremy Huckins, Jason Holden, Sarah M. Preum, Colin Depp, Nicholas Jacobson, Mary Czerwinski, Eric Granholm, Andrew T. Campbell","doi":"arxiv-2409.09570","DOIUrl":"https://doi.org/arxiv-2409.09570","url":null,"abstract":"Mental health concerns are prevalent among college students, highlighting the\u0000need for effective interventions that promote self-awareness and holistic\u0000well-being. MindScape pioneers a novel approach to AI-powered journaling by\u0000integrating passively collected behavioral patterns such as conversational\u0000engagement, sleep, and location with Large Language Models (LLMs). This\u0000integration creates a highly personalized and context-aware journaling\u0000experience, enhancing self-awareness and well-being by embedding behavioral\u0000intelligence into AI. We present an 8-week exploratory study with 20 college\u0000students, demonstrating the MindScape app's efficacy in enhancing positive\u0000affect (7%), reducing negative affect (11%), loneliness (6%), and anxiety and\u0000depression, with a significant week-over-week decrease in PHQ-4 scores (-0.25\u0000coefficient), alongside improvements in mindfulness (7%) and self-reflection\u0000(6%). The study highlights the advantages of contextual AI journaling, with\u0000participants particularly appreciating the tailored prompts and insights\u0000provided by the MindScape app. Our analysis also includes a comparison of\u0000responses to AI-driven contextual versus generic prompts, participant feedback\u0000insights, and proposed strategies for leveraging contextual AI journaling to\u0000improve well-being on college campuses. By showcasing the potential of\u0000contextual AI journaling to support mental health, we provide a foundation for\u0000further investigation into the effects of contextual AI journaling on mental\u0000health and well-being.","PeriodicalId":501541,"journal":{"name":"arXiv - CS - Human-Computer Interaction","volume":"15 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142252481","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AACessTalk: Fostering Communication between Minimally Verbal Autistic Children and Parents with Contextual Guidance and Card Recommendation AACessTalk:通过情境引导和卡片推荐,促进语言能力极弱的自闭症儿童与父母之间的交流
Pub Date : 2024-09-15 DOI: arxiv-2409.09641
Dasom Choi, SoHyun Park, Kyungah Lee, Hwajung Hong, Young-Ho Kim
As minimally verbal autistic (MVA) children communicate with parents throughfew words and nonverbal cues, parents often struggle to encourage theirchildren to express subtle emotions and needs and to grasp their nuancedsignals. We present AACessTalk, a tablet-based, AI-mediated communicationsystem that facilitates meaningful exchanges between an MVA child and a parent.AACessTalk provides real-time guides to the parent to engage the child inconversation and, in turn, recommends contextual vocabulary cards to the child.Through a two-week deployment study with 11 MVA child-parent dyads, we examinehow AACessTalk fosters everyday conversation practice and mutual engagement.Our findings show high engagement from all dyads, leading to increasedfrequency of conversation and turn-taking. AACessTalk also encouraged parentsto explore their own interaction strategies and empowered the children to havemore agency in communication. We discuss the implications of designingtechnologies for balanced communication dynamics in parent-MVA childinteraction.
由于患有自闭症的儿童很少用语言和非语言暗示与父母交流,父母往往很难鼓励他们的孩子表达微妙的情感和需求,也很难把握他们的细微信号。我们介绍的 AACessTalk 是一款基于平板电脑的人工智能中介交流系统,它能促进 MVA 儿童与父母之间进行有意义的交流。AACessTalk 为父母提供实时指导,以吸引儿童参与对话,反过来,它还会向儿童推荐语境词汇卡。AACessTalk 还鼓励家长探索自己的互动策略,并增强了儿童在交流中的自主权。我们讨论了在父母与视障儿童互动中设计平衡交流动力技术的意义。
{"title":"AACessTalk: Fostering Communication between Minimally Verbal Autistic Children and Parents with Contextual Guidance and Card Recommendation","authors":"Dasom Choi, SoHyun Park, Kyungah Lee, Hwajung Hong, Young-Ho Kim","doi":"arxiv-2409.09641","DOIUrl":"https://doi.org/arxiv-2409.09641","url":null,"abstract":"As minimally verbal autistic (MVA) children communicate with parents through\u0000few words and nonverbal cues, parents often struggle to encourage their\u0000children to express subtle emotions and needs and to grasp their nuanced\u0000signals. We present AACessTalk, a tablet-based, AI-mediated communication\u0000system that facilitates meaningful exchanges between an MVA child and a parent.\u0000AACessTalk provides real-time guides to the parent to engage the child in\u0000conversation and, in turn, recommends contextual vocabulary cards to the child.\u0000Through a two-week deployment study with 11 MVA child-parent dyads, we examine\u0000how AACessTalk fosters everyday conversation practice and mutual engagement.\u0000Our findings show high engagement from all dyads, leading to increased\u0000frequency of conversation and turn-taking. AACessTalk also encouraged parents\u0000to explore their own interaction strategies and empowered the children to have\u0000more agency in communication. We discuss the implications of designing\u0000technologies for balanced communication dynamics in parent-MVA child\u0000interaction.","PeriodicalId":501541,"journal":{"name":"arXiv - CS - Human-Computer Interaction","volume":"20 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142252479","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spatial-Temporal Mamba Network for EEG-based Motor Imagery Classification 基于脑电图的运动图像分类时空曼巴网络
Pub Date : 2024-09-15 DOI: arxiv-2409.09627
Xiaoxiao Yang, Ziyu Jia
Motor imagery (MI) classification is key for brain-computer interfaces(BCIs). Until recent years, numerous models had been proposed, ranging fromclassical algorithms like Common Spatial Pattern (CSP) to deep learning modelssuch as convolutional neural networks (CNNs) and transformers. However, thesemodels have shown limitations in areas such as generalizability, contextualityand scalability when it comes to effectively extracting the complexspatial-temporal information inherent in electroencephalography (EEG) signals.To address these limitations, we introduce Spatial-Temporal Mamba Network(STMambaNet), an innovative model leveraging the Mamba state spacearchitecture, which excels in processing extended sequences with linearscalability. By incorporating spatial and temporal Mamba encoders, STMambaNeteffectively captures the intricate dynamics in both space and time,significantly enhancing the decoding performance of EEG signals for MIclassification. Experimental results on BCI Competition IV 2a and 2b datasetsdemonstrate STMambaNet's superiority over existing models, establishing it as apowerful tool for advancing MI-based BCIs and improving real-world BCI systems.
运动图像(MI)分类是脑机接口(BCI)的关键。近年来,从通用空间模式(CSP)等经典算法到卷积神经网络(CNN)和变换器等深度学习模型,人们已经提出了许多模型。为了解决这些局限性,我们引入了空间-时间曼巴网络(STMambaNet),这是一种利用曼巴状态空间架构的创新模型,在处理扩展序列时具有出色的线性可扩展性。通过结合空间和时间曼巴编码器,STMambaNet 有效地捕捉了空间和时间的复杂动态,大大提高了用于 MI 分类的脑电信号的解码性能。在 BCI Competition IV 2a 和 2b 数据集上的实验结果证明了 STMambaNet 优于现有模型,使其成为推进基于 MI 的 BCI 和改进实际 BCI 系统的有力工具。
{"title":"Spatial-Temporal Mamba Network for EEG-based Motor Imagery Classification","authors":"Xiaoxiao Yang, Ziyu Jia","doi":"arxiv-2409.09627","DOIUrl":"https://doi.org/arxiv-2409.09627","url":null,"abstract":"Motor imagery (MI) classification is key for brain-computer interfaces\u0000(BCIs). Until recent years, numerous models had been proposed, ranging from\u0000classical algorithms like Common Spatial Pattern (CSP) to deep learning models\u0000such as convolutional neural networks (CNNs) and transformers. However, these\u0000models have shown limitations in areas such as generalizability, contextuality\u0000and scalability when it comes to effectively extracting the complex\u0000spatial-temporal information inherent in electroencephalography (EEG) signals.\u0000To address these limitations, we introduce Spatial-Temporal Mamba Network\u0000(STMambaNet), an innovative model leveraging the Mamba state space\u0000architecture, which excels in processing extended sequences with linear\u0000scalability. By incorporating spatial and temporal Mamba encoders, STMambaNet\u0000effectively captures the intricate dynamics in both space and time,\u0000significantly enhancing the decoding performance of EEG signals for MI\u0000classification. Experimental results on BCI Competition IV 2a and 2b datasets\u0000demonstrate STMambaNet's superiority over existing models, establishing it as a\u0000powerful tool for advancing MI-based BCIs and improving real-world BCI systems.","PeriodicalId":501541,"journal":{"name":"arXiv - CS - Human-Computer Interaction","volume":"9 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142252480","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
arXiv - CS - Human-Computer Interaction
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1