首页 > 最新文献

ASSETS. Annual ACM Conference on Assistive Technologies最新文献

英文 中文
Modeling Accessibility: Characterizing What We Mean by "Accessible". 建模可访问性:描述我们所说的“可访问性”。
Pub Date : 2025-10-01 Epub Date: 2025-10-22 DOI: 10.1145/3663547.3746344
Kelly Avery Mack, Jennifer Mankoff, Heather D Evans, Jesse J Martinez, James Fogarty, Cynthia L Bennett, Aaleyah Lewis, Leah Findlater, Emma J McDonnell

Accessibility research has a broad mandate: use technology to make the world more accessible to disabled people. Yet, as a field, accessibility research lacks a clear characterization of what "accessibility" is. Furthermore, it has been historically limited in who is designed for, focusing on specific types of disability and often failing to consider how disability intersects with other identities. We set out to explicate what it means to make something accessible, grounded in the lived experiences of a diverse group of 25 disabled people. From our empirical findings, we develop a process for modeling accessibility. First, an individual assesses their experience of inaccess, specifically, the type of barrier they face, the technology repertoire they possess, and the contextual factors that shape how they address accessibility barriers. Then, having assessed an access barrier, they perform consequence calculus, weighing all available options to achieve access and deciding upon the option that best matches their priorities. We highlight the situated nature of access; people's identities, contextual factors, repertoires, and priorities all dictate their experience of accessibility.

无障碍研究有一个广泛的任务:利用技术使世界对残疾人更方便。然而,作为一个研究领域,可及性研究缺乏对“可及性”的明确定义。此外,它在历史上一直局限于为谁设计,专注于特定类型的残疾,往往没有考虑残疾如何与其他身份相交。我们以25名不同群体的残疾人的生活经历为基础,着手解释什么是无障碍。根据我们的经验发现,我们开发了一个建模可访问性的过程。首先,个人评估他们的无障碍体验,具体来说,他们面临的障碍类型,他们拥有的技术储备,以及影响他们如何解决无障碍障碍的背景因素。然后,在评估了访问障碍后,他们执行结果演算,权衡所有可用的选项以实现访问,并决定最符合他们优先级的选项。我们强调通道的位置性质;人们的身份、背景因素、技能和优先级都决定了他们对无障碍的体验。
{"title":"Modeling Accessibility: Characterizing What We Mean by \"Accessible\".","authors":"Kelly Avery Mack, Jennifer Mankoff, Heather D Evans, Jesse J Martinez, James Fogarty, Cynthia L Bennett, Aaleyah Lewis, Leah Findlater, Emma J McDonnell","doi":"10.1145/3663547.3746344","DOIUrl":"https://doi.org/10.1145/3663547.3746344","url":null,"abstract":"<p><p>Accessibility research has a broad mandate: use technology to make the world more accessible to disabled people. Yet, as a field, accessibility research lacks a clear characterization of what \"accessibility\" is. Furthermore, it has been historically limited in who is designed for, focusing on specific types of disability and often failing to consider how disability intersects with other identities. We set out to explicate what it means to make something accessible, grounded in the lived experiences of a diverse group of 25 disabled people. From our empirical findings, we develop a process for modeling accessibility. First, an individual assesses their experience of inaccess, specifically, the type of barrier they face, the technology repertoire they possess, and the contextual factors that shape how they address accessibility barriers. Then, having assessed an access barrier, they perform consequence calculus, weighing all available options to achieve access and deciding upon the option that best matches their priorities. We highlight the situated nature of access; people's identities, contextual factors, repertoires, and priorities all dictate their experience of accessibility.</p>","PeriodicalId":72321,"journal":{"name":"ASSETS. Annual ACM Conference on Assistive Technologies","volume":"2025 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12969287/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147438114","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Exploring Collaboration to Center the Deaf Community in Sign Language AI. 探索以手语人工智能为中心的聋人社区的合作。
Pub Date : 2025-01-01 DOI: 10.1145/3663547.3746390
Rie Kamikubo, Abraham Glasser, Alex X Lu, Hal Daumé, Hernisa Kacorri, Danielle Bragg

Sign language processing holds great promise for advancing societal inclusivity, yet it often excludes meaningful participation from the Deaf community, raising ethical and practical concerns about the applicability of AI solutions to their needs. This paper addresses these gaps through two interrelated studies. First, surveys identify differences in priorities and expectations between machine learning (ML) practitioners and Deaf American Sign Language (ASL) signers. Second, paired co-design sessions bring ML and ASL experts together to generate guiding questions that support practices for aligning AI development with community goals. Our findings reveal critical points of friction that reflect deeper systemic and epistemic barriers to effective collaboration. By synthesizing unique and shared insights from both groups, we provide empirically grounded resources to guide collaborative frameworks that promote the agency and expertise of the Deaf community. This research paves actionable pathways toward equitable, community-centered advancements in AI.

手语处理在促进社会包容性方面有着巨大的希望,但它往往排除了聋人社区的有意义的参与,引发了对人工智能解决方案是否适用于他们需求的道德和实际担忧。本文通过两个相互关联的研究来解决这些差距。首先,调查确定了机器学习(ML)从业者和美国聋人手语(ASL)签用者在优先级和期望方面的差异。其次,配对的协同设计会议将ML和ASL专家聚集在一起,提出指标性问题,支持将AI开发与社区目标保持一致的实践。我们的研究结果揭示了摩擦的关键点,这些关键点反映了影响有效合作的更深层次的系统和认知障碍。通过综合两个群体的独特和共享的见解,我们提供了基于经验的资源来指导协作框架,以促进聋人社区的代理和专业知识。这项研究为人工智能的公平、以社区为中心的进步铺平了可行的道路。
{"title":"Exploring Collaboration to Center the Deaf Community in Sign Language AI.","authors":"Rie Kamikubo, Abraham Glasser, Alex X Lu, Hal Daumé, Hernisa Kacorri, Danielle Bragg","doi":"10.1145/3663547.3746390","DOIUrl":"10.1145/3663547.3746390","url":null,"abstract":"<p><p>Sign language processing holds great promise for advancing societal inclusivity, yet it often excludes meaningful participation from the Deaf community, raising ethical and practical concerns about the applicability of AI solutions to their needs. This paper addresses these gaps through two interrelated studies. First, surveys identify differences in priorities and expectations between machine learning (ML) practitioners and Deaf American Sign Language (ASL) signers. Second, paired co-design sessions bring ML and ASL experts together to generate guiding questions that support practices for aligning AI development with community goals. Our findings reveal critical points of friction that reflect deeper systemic and epistemic barriers to effective collaboration. By synthesizing unique and shared insights from both groups, we provide empirically grounded resources to guide collaborative frameworks that promote the agency and expertise of the Deaf community. This research paves actionable pathways toward equitable, community-centered advancements in AI.</p>","PeriodicalId":72321,"journal":{"name":"ASSETS. Annual ACM Conference on Assistive Technologies","volume":"1 ","pages":"1-18"},"PeriodicalIF":0.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12955818/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147357775","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hevelius Report: Visualizing Web-Based Mobility Test Data For Clinical Decision and Learning Support. Hevelius报告:可视化基于网络的移动测试数据,用于临床决策和学习支持。
Pub Date : 2024-10-01 Epub Date: 2024-10-27 DOI: 10.1145/3663548.3688490
Hongjin Lin, Tessa Han, Krzysztof Z Gajos, Anoopum S Gupta

Hevelius, a web-based computer mouse test, measures arm movement and has been shown to accurately evaluate severity for patients with Parkinson's disease and ataxias. A Hevelius session produces 32 numeric features, which may be hard to interpret, especially in time-constrained clinical settings. This work aims to support clinicians (and other stakeholders) in interpreting and connecting Hevelius features to clinical concepts. Through an iterative design process, we developed a visualization tool (Hevelius Report) that (1) abstracts six clinically relevant concepts from 32 features, (2) visualizes patient test results, and compares them to results from healthy controls and other patients, and (3) is an interactive app to meet the specific needs in different usage scenarios. Then, we conducted a preliminary user study through an online interview with three clinicians who were not involved in the project. They expressed interest in using Hevelius Report, especially for identifying subtle changes in their patients' mobility that are hard to capture with existing clinical tests. Future work will integrate the visualization tool into the current clinical workflow of a neurology team and conduct systematic evaluations of the tool's usefulness, usability, and effectiveness. Hevelius Report represents a promising solution for analyzing fine-motor test results and monitoring patients' conditions and progressions.

Hevelius是一项基于网络的计算机鼠标测试,测量手臂运动,并被证明可以准确评估帕金森病和共济失调患者的严重程度。Hevelius会话产生32个数字特征,这可能很难解释,特别是在时间有限的临床环境中。这项工作旨在支持临床医生(和其他利益相关者)在解释和连接Hevelius功能的临床概念。通过迭代设计过程,我们开发了一个可视化工具(Hevelius Report),该工具(1)从32个特征中抽象出6个临床相关概念,(2)将患者检测结果可视化,并将其与健康对照和其他患者的结果进行比较,(3)是一个交互式应用程序,以满足不同使用场景的特定需求。然后,我们通过与三位未参与该项目的临床医生的在线访谈进行了初步的用户研究。他们表达了对使用Hevelius Report的兴趣,特别是在识别现有临床试验难以捕捉的患者活动能力的细微变化方面。未来的工作将把可视化工具整合到神经病学团队当前的临床工作流程中,并对该工具的有用性、可用性和有效性进行系统的评估。Hevelius报告代表了分析精细运动测试结果和监测患者状况和进展的有希望的解决方案。
{"title":"Hevelius Report: Visualizing Web-Based Mobility Test Data For Clinical Decision and Learning Support.","authors":"Hongjin Lin, Tessa Han, Krzysztof Z Gajos, Anoopum S Gupta","doi":"10.1145/3663548.3688490","DOIUrl":"10.1145/3663548.3688490","url":null,"abstract":"<p><p><i>Hevelius</i>, a web-based computer mouse test, measures arm movement and has been shown to accurately evaluate severity for patients with Parkinson's disease and ataxias. A <i>Hevelius</i> session produces 32 numeric features, which may be hard to interpret, especially in time-constrained clinical settings. This work aims to support clinicians (and other stakeholders) in interpreting and connecting <i>Hevelius</i> features to clinical concepts. Through an iterative design process, we developed a visualization tool (<i>Hevelius Report</i>) that (1) abstracts six clinically relevant concepts from 32 features, (2) visualizes patient test results, and compares them to results from healthy controls and other patients, and (3) is an interactive app to meet the specific needs in different usage scenarios. Then, we conducted a preliminary user study through an online interview with three clinicians who were <i>not</i> involved in the project. They expressed interest in using <i>Hevelius Report</i>, especially for identifying subtle changes in their patients' mobility that are hard to capture with existing clinical tests. Future work will integrate the visualization tool into the current clinical workflow of a neurology team and conduct systematic evaluations of the tool's usefulness, usability, and effectiveness. <i>Hevelius Report</i> represents a promising solution for analyzing fine-motor test results and monitoring patients' conditions and progressions.</p>","PeriodicalId":72321,"journal":{"name":"ASSETS. Annual ACM Conference on Assistive Technologies","volume":"2024 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12239997/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144602405","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Exploring Videoconferencing for Older Adults with Cognitive Concerns Using a Dramaturgical Lens. 用戏剧镜头探索有认知问题的老年人视频会议。
Pub Date : 2024-01-01 DOI: 10.1145/3663548.3675647
Ruipu Hu, Ge Gao, Amanda Lazar

While videoconferencing is a promising technology, it may present unique challenges and barriers for older adults with cognitive concerns. This paper presents a deconstructed view of videoconferencing technology use using a sociological dramaturgical framework developed by Erving Goffman. Our study recruited 17 older adults with varying cognitive concerns, employing technology discussion groups, interviews, and observations to gather data. Through a reflexive thematic analysis, we explore videoconferencing use among older adults with cognitive concerns, focusing on three major areas: the "performances and roles" where users adapt to new roles through videoconferencing; the "backstage," which involves the physical and logistical setup; and the "frontstage," where people communicate through audio and visual channels to present a desired impression. Our discussion generates insights into how deconstructing these elements can inform more meaningful and accessible HCI design.

虽然视频会议是一项很有前途的技术,但它可能会给有认知问题的老年人带来独特的挑战和障碍。本文采用欧文·戈夫曼(Erving Goffman)提出的社会学戏剧框架,对视频会议技术的使用进行解构。我们的研究招募了17名有不同认知问题的老年人,采用技术讨论小组、访谈和观察来收集数据。通过反身性专题分析,我们探讨了具有认知问题的老年人对视频会议的使用,重点关注三个主要领域:用户通过视频会议适应新角色的“表演和角色”;“后台”,包括物理和后勤设置;还有“前台”,人们通过音频和视觉渠道进行交流,以呈现理想的印象。我们的讨论产生了关于如何解构这些元素以提供更有意义和更易于访问的HCI设计的见解。
{"title":"Exploring Videoconferencing for Older Adults with Cognitive Concerns Using a Dramaturgical Lens.","authors":"Ruipu Hu, Ge Gao, Amanda Lazar","doi":"10.1145/3663548.3675647","DOIUrl":"10.1145/3663548.3675647","url":null,"abstract":"<p><p>While videoconferencing is a promising technology, it may present unique challenges and barriers for older adults with cognitive concerns. This paper presents a deconstructed view of videoconferencing technology use using a sociological dramaturgical framework developed by Erving Goffman. Our study recruited 17 older adults with varying cognitive concerns, employing technology discussion groups, interviews, and observations to gather data. Through a reflexive thematic analysis, we explore videoconferencing use among older adults with cognitive concerns, focusing on three major areas: the \"performances and roles\" where users adapt to new roles through videoconferencing; the \"backstage,\" which involves the physical and logistical setup; and the \"frontstage,\" where people communicate through audio and visual channels to present a desired impression. Our discussion generates insights into how deconstructing these elements can inform more meaningful and accessible HCI design.</p>","PeriodicalId":72321,"journal":{"name":"ASSETS. Annual ACM Conference on Assistive Technologies","volume":"2024 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12188971/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144499621","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AccessShare: Co-designing Data Access and Sharing with Blind People. AccessShare:与盲人共同设计数据访问和共享。
Pub Date : 2024-01-01 DOI: 10.1145/3663548.3675612
Rie Kamikubo, Farnaz Zamiri Zeraati, Kyungjun Lee, Hernisa Kacorri

Blind people are often called to contribute image data to datasets for AI innovation with the hope for future accessibility and inclusion. Yet, the visual inspection of the contributed images is inaccessible. To this day, we lack mechanisms for data inspection and control that are accessible to the blind community. To address this gap, we engage 10 blind participants in a scenario where they wear smartglasses and collect image data using an AI-infused application in their homes. We also engineer a design probe, a novel data access interface called AccessShare, and conduct a co-design study to discuss participants' needs, preferences, and ideas on consent, data inspection, and control. Our findings reveal the impact of interactive informed consent and the complementary role of data inspection systems such as AccessShare in facilitating communication between data stewards and blind data contributors. We discuss how key insights can guide future informed consent and data control to promote inclusive and responsible data practices in AI.

盲人经常被要求为人工智能创新的数据集提供图像数据,希望未来的可访问性和包容性。然而,对所提供的图像进行视觉检查是不可访问的。直到今天,我们还缺乏盲人社区可以使用的数据检查和控制机制。为了解决这一差距,我们让10名盲人参与者参与了一个场景,他们戴着智能眼镜,在家中使用人工智能应用程序收集图像数据。我们还设计了一个设计探针,一个名为AccessShare的新型数据访问接口,并进行了一项共同设计研究,以讨论参与者在同意、数据检查和控制方面的需求、偏好和想法。我们的研究结果揭示了交互式知情同意的影响以及AccessShare等数据检查系统在促进数据管理员和盲数据贡献者之间沟通方面的补充作用。我们讨论了关键见解如何指导未来的知情同意和数据控制,以促进人工智能中包容和负责任的数据实践。
{"title":"AccessShare: Co-designing Data Access and Sharing with Blind People.","authors":"Rie Kamikubo, Farnaz Zamiri Zeraati, Kyungjun Lee, Hernisa Kacorri","doi":"10.1145/3663548.3675612","DOIUrl":"10.1145/3663548.3675612","url":null,"abstract":"<p><p>Blind people are often called to contribute image data to datasets for AI innovation with the hope for future accessibility and inclusion. Yet, the visual inspection of the contributed images is inaccessible. To this day, we lack mechanisms for data inspection and control that are accessible to the blind community. To address this gap, we engage 10 blind participants in a scenario where they wear smartglasses and collect image data using an AI-infused application in their homes. We also engineer a design probe, a novel data access interface called AccessShare, and conduct a co-design study to discuss participants' needs, preferences, and ideas on consent, data inspection, and control. Our findings reveal the impact of interactive informed consent and the complementary role of data inspection systems such as AccessShare in facilitating communication between data stewards and blind data contributors. We discuss how key insights can guide future informed consent and data control to promote inclusive and responsible data practices in AI.</p>","PeriodicalId":72321,"journal":{"name":"ASSETS. Annual ACM Conference on Assistive Technologies","volume":"4 ","pages":"1-16"},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12188854/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144499622","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Understanding How Blind Users Handle Object Recognition Errors: Strategies and Challenges. 理解盲人用户如何处理对象识别错误:策略和挑战。
Pub Date : 2024-01-01 DOI: 10.1145/3663548.3675635
Jonggi Hong, Hernisa Kacorri

Object recognition technologies hold the potential to support blind and low-vision people in navigating the world around them. However, the gap between benchmark performances and practical usability remains a significant challenge. This paper presents a study aimed at understanding blind users' interaction with object recognition systems for identifying and avoiding errors. Leveraging a pre-existing object recognition system, URCam, fine-tuned for our experiment, we conducted a user study involving 12 blind and low-vision participants. Through in-depth interviews and hands-on error identification tasks, we gained insights into users' experiences, challenges, and strategies for identifying errors in camera-based assistive technologies and object recognition systems. During interviews, many participants preferred independent error review, while expressing apprehension toward misrecognitions. In the error identification task, participants varied viewpoints, backgrounds, and object sizes in their images to avoid and overcome errors. Even after repeating the task, participants identified only half of the errors, and the proportion of errors identified did not significantly differ from their first attempts. Based on these insights, we offer implications for designing accessible interfaces tailored to the needs of blind and low-vision users in identifying object recognition errors.

物体识别技术有潜力支持盲人和低视力人群在他们周围的世界中导航。然而,基准性能和实际可用性之间的差距仍然是一个重大挑战。本文提出了一项研究,旨在了解盲人用户与目标识别系统的交互,以识别和避免错误。利用预先存在的物体识别系统URCam,为我们的实验进行了微调,我们进行了一项涉及12名盲人和低视力参与者的用户研究。通过深入访谈和实践错误识别任务,我们深入了解了用户在基于相机的辅助技术和对象识别系统中识别错误的经验、挑战和策略。在采访中,许多参与者倾向于独立的错误审查,同时对错误认识表示担忧。在错误识别任务中,参与者通过改变图像中的视角、背景和物体大小来避免和克服错误。即使在重复了任务之后,参与者也只发现了一半的错误,而且发现的错误比例与第一次尝试时没有显著差异。基于这些见解,我们为设计适合盲人和低视力用户识别物体识别错误需求的可访问界面提供了启示。
{"title":"Understanding How Blind Users Handle Object Recognition Errors: Strategies and Challenges.","authors":"Jonggi Hong, Hernisa Kacorri","doi":"10.1145/3663548.3675635","DOIUrl":"10.1145/3663548.3675635","url":null,"abstract":"<p><p>Object recognition technologies hold the potential to support blind and low-vision people in navigating the world around them. However, the gap between benchmark performances and practical usability remains a significant challenge. This paper presents a study aimed at understanding blind users' interaction with object recognition systems for identifying and avoiding errors. Leveraging a pre-existing object recognition system, URCam, fine-tuned for our experiment, we conducted a user study involving 12 blind and low-vision participants. Through in-depth interviews and hands-on error identification tasks, we gained insights into users' experiences, challenges, and strategies for identifying errors in camera-based assistive technologies and object recognition systems. During interviews, many participants preferred independent error review, while expressing apprehension toward misrecognitions. In the error identification task, participants varied viewpoints, backgrounds, and object sizes in their images to avoid and overcome errors. Even after repeating the task, participants identified only half of the errors, and the proportion of errors identified did not significantly differ from their first attempts. Based on these insights, we offer implications for designing accessible interfaces tailored to the needs of blind and low-vision users in identifying object recognition errors.</p>","PeriodicalId":72321,"journal":{"name":"ASSETS. Annual ACM Conference on Assistive Technologies","volume":"2024 ","pages":"1-15"},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11872236/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143544836","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enabling Uniform Computer Interaction Experience for Blind Users through Large Language Models. 通过大型语言模型实现盲人用户统一的计算机交互体验。
Pub Date : 2024-01-01 Epub Date: 2024-10-27 DOI: 10.1145/3663548.3675605
Satwik Ram Kodandaram, Utku Uckun, Xiaojun Bi, I V Ramakrishnan, Vikas Ashok

Blind individuals, who by necessity depend on screen readers to interact with computers, face considerable challenges in navigating the diverse and complex graphical user interfaces of different computer applications. The heterogeneity of various application interfaces often requires blind users to remember different keyboard combinations and navigation methods to use each application effectively. To alleviate this significant interaction burden imposed by heterogeneous application interfaces, we present Savant, a novel assistive technology powered by large language models (LLMs) that allows blind screen reader users to interact uniformly with any application interface through natural language. Novelly, Savant can automate a series of tedious screen reader actions on the control elements of the application when prompted by a natural language command from the user. These commands can be flexible in the sense that the user is not strictly required to specify the exact names of the control elements in the command. A user study evaluation of Savant with 11 blind participants demonstrated significant improvements in interaction efficiency and usability compared to current practices.

盲人必须依靠屏幕阅读器与计算机交互,他们在浏览不同计算机应用程序的多样化和复杂的图形用户界面时面临相当大的挑战。各种应用程序接口的异构性通常要求盲人用户记住不同的键盘组合和导航方法,以便有效地使用每个应用程序。为了减轻异构应用程序接口带来的交互负担,我们提出了Savant,这是一种由大型语言模型(llm)提供支持的新型辅助技术,它允许盲人屏幕阅读器用户通过自然语言与任何应用程序接口进行统一的交互。新颖的是,当用户发出自然语言命令提示时,Savant可以在应用程序的控制元素上自动执行一系列繁琐的屏幕阅读器操作。这些命令可以很灵活,因为用户不需要在命令中严格指定控制元素的确切名称。一项由11名盲人参与的Savant用户研究评估显示,与目前的做法相比,在交互效率和可用性方面有了显著的改进。
{"title":"Enabling Uniform Computer Interaction Experience for Blind Users through Large Language Models.","authors":"Satwik Ram Kodandaram, Utku Uckun, Xiaojun Bi, I V Ramakrishnan, Vikas Ashok","doi":"10.1145/3663548.3675605","DOIUrl":"10.1145/3663548.3675605","url":null,"abstract":"<p><p>Blind individuals, who by necessity depend on screen readers to interact with computers, face considerable challenges in navigating the diverse and complex graphical user interfaces of different computer applications. The heterogeneity of various application interfaces often requires blind users to remember different keyboard combinations and navigation methods to use each application effectively. To alleviate this significant interaction burden imposed by heterogeneous application interfaces, we present Savant, a novel assistive technology powered by large language models (LLMs) that allows blind screen reader users to interact uniformly with any application interface through natural language. Novelly, Savant can automate a series of tedious screen reader actions on the control elements of the application when prompted by a natural language command from the user. These commands can be flexible in the sense that the user is not strictly required to specify the exact names of the control elements in the command. A user study evaluation of Savant with 11 blind participants demonstrated significant improvements in interaction efficiency and usability compared to current practices.</p>","PeriodicalId":72321,"journal":{"name":"ASSETS. Annual ACM Conference on Assistive Technologies","volume":"2024 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11707650/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142959794","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Information Wayfinding of Screen Reader Users: Five Personas to Expand Conceptualizations of User Experiences. 屏幕阅读器用户的信息寻路:扩展用户体验概念的五个角色。
J Bern Jordan, Victoria Van Hyning, Mason A Jones, Rachael Bradley Montgomery, Elizabeth Bottner, Evan Tansil

Screen readers are important assistive technologies for blind people, but they are complex and can be challenging to use effectively. Over the course of several studies with screen reader users, the authors have found wide variations and sometimes surprising differences in people's skills, preferences, navigation, and troubleshooting approaches when using screen readers. These differences may not always be considered in research and development. To help address this shortcoming, we have developed five user personas describing a range of screen reader experiences.

屏幕阅读器对盲人来说是重要的辅助技术,但它们很复杂,很难有效地使用。在对屏幕阅读器用户进行的几项研究中,作者发现,在使用屏幕阅读器时,人们的技能、偏好、导航和故障排除方法存在很大的差异,有时甚至令人惊讶。这些差异在研究和开发中可能并不总是被考虑到。为了帮助解决这个缺点,我们开发了五个用户角色来描述一系列屏幕阅读器的体验。
{"title":"Information Wayfinding of Screen Reader Users: Five Personas to Expand Conceptualizations of User Experiences.","authors":"J Bern Jordan, Victoria Van Hyning, Mason A Jones, Rachael Bradley Montgomery, Elizabeth Bottner, Evan Tansil","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Screen readers are important assistive technologies for blind people, but they are complex and can be challenging to use effectively. Over the course of several studies with screen reader users, the authors have found wide variations and sometimes surprising differences in people's skills, preferences, navigation, and troubleshooting approaches when using screen readers. These differences may not always be considered in research and development. To help address this shortcoming, we have developed five user personas describing a range of screen reader experiences.</p>","PeriodicalId":72321,"journal":{"name":"ASSETS. Annual ACM Conference on Assistive Technologies","volume":"2024 47","pages":"1-7"},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11872227/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143544839","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Screen Magnification for Readers with Low Vision: A Study on Usability and Performance. 低视力读者的屏幕放大功能:可用性和性能研究
Pub Date : 2023-10-01 Epub Date: 2023-10-22
Meini Tang, Roberto Manduchi, Susana Chung, Raquel Prado

We present a study with 20 participants with low vision who operated two types of screen magnification (lens and full) on a laptop computer to read two types of document (text and web page). Our purposes were to comparatively assess the two magnification modalities, and to obtain some insight into how people with low vision use the mouse to control the center of magnification. These observations may inform the design of systems for the automatic control of the center of magnification. Our results show that there were no significant differences in reading performances or in subjective preferences between the two magnification modes. However, when using the lens mode, our participants adopted more consistent and uniform mouse motion patterns, while longer and more frequent pauses and shorter overall path lengths were measured using the full mode. Analysis of the distribution of gaze points (as measured by a gaze tracker) using the full mode shows that, when reading a text document, most participants preferred to move the area of interest to a specific region of the screen.

我们对 20 名低视能参与者进行了一项研究,他们在笔记本电脑上操作两种屏幕放大方式(镜头放大和完全放大)来阅读两种类型的文档(文本和网页)。我们的目的是对两种放大模式进行比较评估,并深入了解低视力者如何使用鼠标控制放大中心。这些观察结果可为设计自动控制放大中心的系统提供参考。我们的研究结果表明,两种放大模式在阅读效果和主观偏好上没有明显差异。不过,在使用透镜模式时,我们的参与者采用了更一致、更均匀的鼠标运动模式,而在使用完整模式时,他们的停顿时间更长、更频繁,总体路径长度也更短。对使用全屏模式时的注视点分布(由注视跟踪器测量)进行的分析表明,在阅读文本文档时,大多数参与者更倾向于将感兴趣的区域移动到屏幕的特定区域。
{"title":"Screen Magnification for Readers with Low Vision: A Study on Usability and Performance.","authors":"Meini Tang, Roberto Manduchi, Susana Chung, Raquel Prado","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>We present a study with 20 participants with low vision who operated two types of screen magnification (lens and full) on a laptop computer to read two types of document (text and web page). Our purposes were to comparatively assess the two magnification modalities, and to obtain some insight into how people with low vision use the mouse to control the center of magnification. These observations may inform the design of systems for the automatic control of the center of magnification. Our results show that there were no significant differences in reading performances or in subjective preferences between the two magnification modes. However, when using the lens mode, our participants adopted more consistent and uniform mouse motion patterns, while longer and more frequent pauses and shorter overall path lengths were measured using the full mode. Analysis of the distribution of gaze points (as measured by a gaze tracker) using the full mode shows that, when reading a text document, most participants preferred to move the area of interest to a specific region of the screen.</p>","PeriodicalId":72321,"journal":{"name":"ASSETS. Annual ACM Conference on Assistive Technologies","volume":"2023 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10923554/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140095279","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Blind Users Accessing Their Training Images in Teachable Object Recognizers. 盲人用户在可教物体识别器中访问自己的训练图像
Pub Date : 2022-10-01 Epub Date: 2022-10-22 DOI: 10.1145/3517428.3544824
Jonggi Hong, Jaina Gandhi, Ernest Essuah Mensah, Farnaz Zamiri Zeraati, Ebrima Haddy Jarjue, Kyungjun Lee, Hernisa Kacorri

Teachable object recognizers provide a solution for a very practical need for blind people - instance level object recognition. They assume one can visually inspect the photos they provide for training, a critical and inaccessible step for those who are blind. In this work, we engineer data descriptors that address this challenge. They indicate in real time whether the object in the photo is cropped or too small, a hand is included, the photos is blurred, and how much photos vary from each other. Our descriptors are built into open source testbed iOS app, called MYCam. In a remote user study in (N = 12) blind participants' homes, we show how descriptors, even when error-prone, support experimentation and have a positive impact in the quality of training set that can translate to model performance though this gain is not uniform. Participants found the app simple to use indicating that they could effectively train it and that the descriptors were useful. However, many found the training being tedious, opening discussions around the need for balance between information, time, and cognitive load.

可教物体识别器为盲人的一个非常实际的需求提供了解决方案--实例级物体识别。它们假定人们可以目测它们提供的用于训练的照片,而对于盲人来说,这是一个关键且难以接近的步骤。在这项工作中,我们设计了数据描述符来应对这一挑战。它们能实时显示照片中的物体是否被裁剪或太小、是否包含一只手、照片是否模糊以及照片之间的差异程度。我们的描述符内置于名为 MYCam 的开源测试平台 iOS 应用程序中。在盲人参与者家中进行的远程用户研究(N = 12)中,我们展示了描述符(即使容易出错)是如何支持实验并对训练集的质量产生积极影响的,这种影响可以转化为模型性能,尽管这种增益并不一致。参与者认为该应用简单易用,表明他们可以有效地对其进行训练,而且描述符也很有用。不过,许多人认为训练很枯燥乏味,因此开始讨论在信息、时间和认知负荷之间保持平衡的必要性。
{"title":"Blind Users Accessing Their Training Images in Teachable Object Recognizers.","authors":"Jonggi Hong, Jaina Gandhi, Ernest Essuah Mensah, Farnaz Zamiri Zeraati, Ebrima Haddy Jarjue, Kyungjun Lee, Hernisa Kacorri","doi":"10.1145/3517428.3544824","DOIUrl":"10.1145/3517428.3544824","url":null,"abstract":"<p><p>Teachable object recognizers provide a solution for a very practical need for blind people - instance level object recognition. They assume one can visually inspect the photos they provide for training, a critical and inaccessible step for those who are blind. In this work, we engineer data descriptors that address this challenge. They indicate in real time whether the object in the photo is cropped or too small, a hand is included, the photos is blurred, and how much photos vary from each other. Our descriptors are built into open source testbed iOS app, called MYCam. In a remote user study in (<i>N</i> = 12) blind participants' homes, we show how descriptors, even when error-prone, support experimentation and have a positive impact in the quality of training set that can translate to model performance though this gain is not uniform. Participants found the app simple to use indicating that they could effectively train it and that the descriptors were useful. However, many found the training being tedious, opening discussions around the need for balance between information, time, and cognitive load.</p>","PeriodicalId":72321,"journal":{"name":"ASSETS. Annual ACM Conference on Assistive Technologies","volume":"2022 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10008526/pdf/nihms-1869981.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9111608","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
ASSETS. Annual ACM Conference on Assistive Technologies
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1