首页 > 最新文献

ASSETS. Annual ACM Conference on Assistive Technologies最新文献

英文 中文
Hevelius Report: Visualizing Web-Based Mobility Test Data For Clinical Decision and Learning Support. Hevelius报告:可视化基于网络的移动测试数据,用于临床决策和学习支持。
Pub Date : 2024-10-01 Epub Date: 2024-10-27 DOI: 10.1145/3663548.3688490
Hongjin Lin, Tessa Han, Krzysztof Z Gajos, Anoopum S Gupta

Hevelius, a web-based computer mouse test, measures arm movement and has been shown to accurately evaluate severity for patients with Parkinson's disease and ataxias. A Hevelius session produces 32 numeric features, which may be hard to interpret, especially in time-constrained clinical settings. This work aims to support clinicians (and other stakeholders) in interpreting and connecting Hevelius features to clinical concepts. Through an iterative design process, we developed a visualization tool (Hevelius Report) that (1) abstracts six clinically relevant concepts from 32 features, (2) visualizes patient test results, and compares them to results from healthy controls and other patients, and (3) is an interactive app to meet the specific needs in different usage scenarios. Then, we conducted a preliminary user study through an online interview with three clinicians who were not involved in the project. They expressed interest in using Hevelius Report, especially for identifying subtle changes in their patients' mobility that are hard to capture with existing clinical tests. Future work will integrate the visualization tool into the current clinical workflow of a neurology team and conduct systematic evaluations of the tool's usefulness, usability, and effectiveness. Hevelius Report represents a promising solution for analyzing fine-motor test results and monitoring patients' conditions and progressions.

Hevelius是一项基于网络的计算机鼠标测试,测量手臂运动,并被证明可以准确评估帕金森病和共济失调患者的严重程度。Hevelius会话产生32个数字特征,这可能很难解释,特别是在时间有限的临床环境中。这项工作旨在支持临床医生(和其他利益相关者)在解释和连接Hevelius功能的临床概念。通过迭代设计过程,我们开发了一个可视化工具(Hevelius Report),该工具(1)从32个特征中抽象出6个临床相关概念,(2)将患者检测结果可视化,并将其与健康对照和其他患者的结果进行比较,(3)是一个交互式应用程序,以满足不同使用场景的特定需求。然后,我们通过与三位未参与该项目的临床医生的在线访谈进行了初步的用户研究。他们表达了对使用Hevelius Report的兴趣,特别是在识别现有临床试验难以捕捉的患者活动能力的细微变化方面。未来的工作将把可视化工具整合到神经病学团队当前的临床工作流程中,并对该工具的有用性、可用性和有效性进行系统的评估。Hevelius报告代表了分析精细运动测试结果和监测患者状况和进展的有希望的解决方案。
{"title":"Hevelius Report: Visualizing Web-Based Mobility Test Data For Clinical Decision and Learning Support.","authors":"Hongjin Lin, Tessa Han, Krzysztof Z Gajos, Anoopum S Gupta","doi":"10.1145/3663548.3688490","DOIUrl":"10.1145/3663548.3688490","url":null,"abstract":"<p><p><i>Hevelius</i>, a web-based computer mouse test, measures arm movement and has been shown to accurately evaluate severity for patients with Parkinson's disease and ataxias. A <i>Hevelius</i> session produces 32 numeric features, which may be hard to interpret, especially in time-constrained clinical settings. This work aims to support clinicians (and other stakeholders) in interpreting and connecting <i>Hevelius</i> features to clinical concepts. Through an iterative design process, we developed a visualization tool (<i>Hevelius Report</i>) that (1) abstracts six clinically relevant concepts from 32 features, (2) visualizes patient test results, and compares them to results from healthy controls and other patients, and (3) is an interactive app to meet the specific needs in different usage scenarios. Then, we conducted a preliminary user study through an online interview with three clinicians who were <i>not</i> involved in the project. They expressed interest in using <i>Hevelius Report</i>, especially for identifying subtle changes in their patients' mobility that are hard to capture with existing clinical tests. Future work will integrate the visualization tool into the current clinical workflow of a neurology team and conduct systematic evaluations of the tool's usefulness, usability, and effectiveness. <i>Hevelius Report</i> represents a promising solution for analyzing fine-motor test results and monitoring patients' conditions and progressions.</p>","PeriodicalId":72321,"journal":{"name":"ASSETS. Annual ACM Conference on Assistive Technologies","volume":"2024 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12239997/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144602405","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Exploring Videoconferencing for Older Adults with Cognitive Concerns Using a Dramaturgical Lens. 用戏剧镜头探索有认知问题的老年人视频会议。
Pub Date : 2024-01-01 DOI: 10.1145/3663548.3675647
Ruipu Hu, Ge Gao, Amanda Lazar

While videoconferencing is a promising technology, it may present unique challenges and barriers for older adults with cognitive concerns. This paper presents a deconstructed view of videoconferencing technology use using a sociological dramaturgical framework developed by Erving Goffman. Our study recruited 17 older adults with varying cognitive concerns, employing technology discussion groups, interviews, and observations to gather data. Through a reflexive thematic analysis, we explore videoconferencing use among older adults with cognitive concerns, focusing on three major areas: the "performances and roles" where users adapt to new roles through videoconferencing; the "backstage," which involves the physical and logistical setup; and the "frontstage," where people communicate through audio and visual channels to present a desired impression. Our discussion generates insights into how deconstructing these elements can inform more meaningful and accessible HCI design.

虽然视频会议是一项很有前途的技术,但它可能会给有认知问题的老年人带来独特的挑战和障碍。本文采用欧文·戈夫曼(Erving Goffman)提出的社会学戏剧框架,对视频会议技术的使用进行解构。我们的研究招募了17名有不同认知问题的老年人,采用技术讨论小组、访谈和观察来收集数据。通过反身性专题分析,我们探讨了具有认知问题的老年人对视频会议的使用,重点关注三个主要领域:用户通过视频会议适应新角色的“表演和角色”;“后台”,包括物理和后勤设置;还有“前台”,人们通过音频和视觉渠道进行交流,以呈现理想的印象。我们的讨论产生了关于如何解构这些元素以提供更有意义和更易于访问的HCI设计的见解。
{"title":"Exploring Videoconferencing for Older Adults with Cognitive Concerns Using a Dramaturgical Lens.","authors":"Ruipu Hu, Ge Gao, Amanda Lazar","doi":"10.1145/3663548.3675647","DOIUrl":"10.1145/3663548.3675647","url":null,"abstract":"<p><p>While videoconferencing is a promising technology, it may present unique challenges and barriers for older adults with cognitive concerns. This paper presents a deconstructed view of videoconferencing technology use using a sociological dramaturgical framework developed by Erving Goffman. Our study recruited 17 older adults with varying cognitive concerns, employing technology discussion groups, interviews, and observations to gather data. Through a reflexive thematic analysis, we explore videoconferencing use among older adults with cognitive concerns, focusing on three major areas: the \"performances and roles\" where users adapt to new roles through videoconferencing; the \"backstage,\" which involves the physical and logistical setup; and the \"frontstage,\" where people communicate through audio and visual channels to present a desired impression. Our discussion generates insights into how deconstructing these elements can inform more meaningful and accessible HCI design.</p>","PeriodicalId":72321,"journal":{"name":"ASSETS. Annual ACM Conference on Assistive Technologies","volume":"2024 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12188971/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144499621","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AccessShare: Co-designing Data Access and Sharing with Blind People. AccessShare:与盲人共同设计数据访问和共享。
Pub Date : 2024-01-01 DOI: 10.1145/3663548.3675612
Rie Kamikubo, Farnaz Zamiri Zeraati, Kyungjun Lee, Hernisa Kacorri

Blind people are often called to contribute image data to datasets for AI innovation with the hope for future accessibility and inclusion. Yet, the visual inspection of the contributed images is inaccessible. To this day, we lack mechanisms for data inspection and control that are accessible to the blind community. To address this gap, we engage 10 blind participants in a scenario where they wear smartglasses and collect image data using an AI-infused application in their homes. We also engineer a design probe, a novel data access interface called AccessShare, and conduct a co-design study to discuss participants' needs, preferences, and ideas on consent, data inspection, and control. Our findings reveal the impact of interactive informed consent and the complementary role of data inspection systems such as AccessShare in facilitating communication between data stewards and blind data contributors. We discuss how key insights can guide future informed consent and data control to promote inclusive and responsible data practices in AI.

盲人经常被要求为人工智能创新的数据集提供图像数据,希望未来的可访问性和包容性。然而,对所提供的图像进行视觉检查是不可访问的。直到今天,我们还缺乏盲人社区可以使用的数据检查和控制机制。为了解决这一差距,我们让10名盲人参与者参与了一个场景,他们戴着智能眼镜,在家中使用人工智能应用程序收集图像数据。我们还设计了一个设计探针,一个名为AccessShare的新型数据访问接口,并进行了一项共同设计研究,以讨论参与者在同意、数据检查和控制方面的需求、偏好和想法。我们的研究结果揭示了交互式知情同意的影响以及AccessShare等数据检查系统在促进数据管理员和盲数据贡献者之间沟通方面的补充作用。我们讨论了关键见解如何指导未来的知情同意和数据控制,以促进人工智能中包容和负责任的数据实践。
{"title":"AccessShare: Co-designing Data Access and Sharing with Blind People.","authors":"Rie Kamikubo, Farnaz Zamiri Zeraati, Kyungjun Lee, Hernisa Kacorri","doi":"10.1145/3663548.3675612","DOIUrl":"10.1145/3663548.3675612","url":null,"abstract":"<p><p>Blind people are often called to contribute image data to datasets for AI innovation with the hope for future accessibility and inclusion. Yet, the visual inspection of the contributed images is inaccessible. To this day, we lack mechanisms for data inspection and control that are accessible to the blind community. To address this gap, we engage 10 blind participants in a scenario where they wear smartglasses and collect image data using an AI-infused application in their homes. We also engineer a design probe, a novel data access interface called AccessShare, and conduct a co-design study to discuss participants' needs, preferences, and ideas on consent, data inspection, and control. Our findings reveal the impact of interactive informed consent and the complementary role of data inspection systems such as AccessShare in facilitating communication between data stewards and blind data contributors. We discuss how key insights can guide future informed consent and data control to promote inclusive and responsible data practices in AI.</p>","PeriodicalId":72321,"journal":{"name":"ASSETS. Annual ACM Conference on Assistive Technologies","volume":"4 ","pages":"1-16"},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12188854/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144499622","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Understanding How Blind Users Handle Object Recognition Errors: Strategies and Challenges. 理解盲人用户如何处理对象识别错误:策略和挑战。
Pub Date : 2024-01-01 DOI: 10.1145/3663548.3675635
Jonggi Hong, Hernisa Kacorri

Object recognition technologies hold the potential to support blind and low-vision people in navigating the world around them. However, the gap between benchmark performances and practical usability remains a significant challenge. This paper presents a study aimed at understanding blind users' interaction with object recognition systems for identifying and avoiding errors. Leveraging a pre-existing object recognition system, URCam, fine-tuned for our experiment, we conducted a user study involving 12 blind and low-vision participants. Through in-depth interviews and hands-on error identification tasks, we gained insights into users' experiences, challenges, and strategies for identifying errors in camera-based assistive technologies and object recognition systems. During interviews, many participants preferred independent error review, while expressing apprehension toward misrecognitions. In the error identification task, participants varied viewpoints, backgrounds, and object sizes in their images to avoid and overcome errors. Even after repeating the task, participants identified only half of the errors, and the proportion of errors identified did not significantly differ from their first attempts. Based on these insights, we offer implications for designing accessible interfaces tailored to the needs of blind and low-vision users in identifying object recognition errors.

物体识别技术有潜力支持盲人和低视力人群在他们周围的世界中导航。然而,基准性能和实际可用性之间的差距仍然是一个重大挑战。本文提出了一项研究,旨在了解盲人用户与目标识别系统的交互,以识别和避免错误。利用预先存在的物体识别系统URCam,为我们的实验进行了微调,我们进行了一项涉及12名盲人和低视力参与者的用户研究。通过深入访谈和实践错误识别任务,我们深入了解了用户在基于相机的辅助技术和对象识别系统中识别错误的经验、挑战和策略。在采访中,许多参与者倾向于独立的错误审查,同时对错误认识表示担忧。在错误识别任务中,参与者通过改变图像中的视角、背景和物体大小来避免和克服错误。即使在重复了任务之后,参与者也只发现了一半的错误,而且发现的错误比例与第一次尝试时没有显著差异。基于这些见解,我们为设计适合盲人和低视力用户识别物体识别错误需求的可访问界面提供了启示。
{"title":"Understanding How Blind Users Handle Object Recognition Errors: Strategies and Challenges.","authors":"Jonggi Hong, Hernisa Kacorri","doi":"10.1145/3663548.3675635","DOIUrl":"10.1145/3663548.3675635","url":null,"abstract":"<p><p>Object recognition technologies hold the potential to support blind and low-vision people in navigating the world around them. However, the gap between benchmark performances and practical usability remains a significant challenge. This paper presents a study aimed at understanding blind users' interaction with object recognition systems for identifying and avoiding errors. Leveraging a pre-existing object recognition system, URCam, fine-tuned for our experiment, we conducted a user study involving 12 blind and low-vision participants. Through in-depth interviews and hands-on error identification tasks, we gained insights into users' experiences, challenges, and strategies for identifying errors in camera-based assistive technologies and object recognition systems. During interviews, many participants preferred independent error review, while expressing apprehension toward misrecognitions. In the error identification task, participants varied viewpoints, backgrounds, and object sizes in their images to avoid and overcome errors. Even after repeating the task, participants identified only half of the errors, and the proportion of errors identified did not significantly differ from their first attempts. Based on these insights, we offer implications for designing accessible interfaces tailored to the needs of blind and low-vision users in identifying object recognition errors.</p>","PeriodicalId":72321,"journal":{"name":"ASSETS. Annual ACM Conference on Assistive Technologies","volume":"2024 ","pages":"1-15"},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11872236/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143544836","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enabling Uniform Computer Interaction Experience for Blind Users through Large Language Models. 通过大型语言模型实现盲人用户统一的计算机交互体验。
Pub Date : 2024-01-01 Epub Date: 2024-10-27 DOI: 10.1145/3663548.3675605
Satwik Ram Kodandaram, Utku Uckun, Xiaojun Bi, I V Ramakrishnan, Vikas Ashok

Blind individuals, who by necessity depend on screen readers to interact with computers, face considerable challenges in navigating the diverse and complex graphical user interfaces of different computer applications. The heterogeneity of various application interfaces often requires blind users to remember different keyboard combinations and navigation methods to use each application effectively. To alleviate this significant interaction burden imposed by heterogeneous application interfaces, we present Savant, a novel assistive technology powered by large language models (LLMs) that allows blind screen reader users to interact uniformly with any application interface through natural language. Novelly, Savant can automate a series of tedious screen reader actions on the control elements of the application when prompted by a natural language command from the user. These commands can be flexible in the sense that the user is not strictly required to specify the exact names of the control elements in the command. A user study evaluation of Savant with 11 blind participants demonstrated significant improvements in interaction efficiency and usability compared to current practices.

盲人必须依靠屏幕阅读器与计算机交互,他们在浏览不同计算机应用程序的多样化和复杂的图形用户界面时面临相当大的挑战。各种应用程序接口的异构性通常要求盲人用户记住不同的键盘组合和导航方法,以便有效地使用每个应用程序。为了减轻异构应用程序接口带来的交互负担,我们提出了Savant,这是一种由大型语言模型(llm)提供支持的新型辅助技术,它允许盲人屏幕阅读器用户通过自然语言与任何应用程序接口进行统一的交互。新颖的是,当用户发出自然语言命令提示时,Savant可以在应用程序的控制元素上自动执行一系列繁琐的屏幕阅读器操作。这些命令可以很灵活,因为用户不需要在命令中严格指定控制元素的确切名称。一项由11名盲人参与的Savant用户研究评估显示,与目前的做法相比,在交互效率和可用性方面有了显著的改进。
{"title":"Enabling Uniform Computer Interaction Experience for Blind Users through Large Language Models.","authors":"Satwik Ram Kodandaram, Utku Uckun, Xiaojun Bi, I V Ramakrishnan, Vikas Ashok","doi":"10.1145/3663548.3675605","DOIUrl":"10.1145/3663548.3675605","url":null,"abstract":"<p><p>Blind individuals, who by necessity depend on screen readers to interact with computers, face considerable challenges in navigating the diverse and complex graphical user interfaces of different computer applications. The heterogeneity of various application interfaces often requires blind users to remember different keyboard combinations and navigation methods to use each application effectively. To alleviate this significant interaction burden imposed by heterogeneous application interfaces, we present Savant, a novel assistive technology powered by large language models (LLMs) that allows blind screen reader users to interact uniformly with any application interface through natural language. Novelly, Savant can automate a series of tedious screen reader actions on the control elements of the application when prompted by a natural language command from the user. These commands can be flexible in the sense that the user is not strictly required to specify the exact names of the control elements in the command. A user study evaluation of Savant with 11 blind participants demonstrated significant improvements in interaction efficiency and usability compared to current practices.</p>","PeriodicalId":72321,"journal":{"name":"ASSETS. Annual ACM Conference on Assistive Technologies","volume":"2024 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11707650/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142959794","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Information Wayfinding of Screen Reader Users: Five Personas to Expand Conceptualizations of User Experiences. 屏幕阅读器用户的信息寻路:扩展用户体验概念的五个角色。
J Bern Jordan, Victoria Van Hyning, Mason A Jones, Rachael Bradley Montgomery, Elizabeth Bottner, Evan Tansil

Screen readers are important assistive technologies for blind people, but they are complex and can be challenging to use effectively. Over the course of several studies with screen reader users, the authors have found wide variations and sometimes surprising differences in people's skills, preferences, navigation, and troubleshooting approaches when using screen readers. These differences may not always be considered in research and development. To help address this shortcoming, we have developed five user personas describing a range of screen reader experiences.

屏幕阅读器对盲人来说是重要的辅助技术,但它们很复杂,很难有效地使用。在对屏幕阅读器用户进行的几项研究中,作者发现,在使用屏幕阅读器时,人们的技能、偏好、导航和故障排除方法存在很大的差异,有时甚至令人惊讶。这些差异在研究和开发中可能并不总是被考虑到。为了帮助解决这个缺点,我们开发了五个用户角色来描述一系列屏幕阅读器的体验。
{"title":"Information Wayfinding of Screen Reader Users: Five Personas to Expand Conceptualizations of User Experiences.","authors":"J Bern Jordan, Victoria Van Hyning, Mason A Jones, Rachael Bradley Montgomery, Elizabeth Bottner, Evan Tansil","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Screen readers are important assistive technologies for blind people, but they are complex and can be challenging to use effectively. Over the course of several studies with screen reader users, the authors have found wide variations and sometimes surprising differences in people's skills, preferences, navigation, and troubleshooting approaches when using screen readers. These differences may not always be considered in research and development. To help address this shortcoming, we have developed five user personas describing a range of screen reader experiences.</p>","PeriodicalId":72321,"journal":{"name":"ASSETS. Annual ACM Conference on Assistive Technologies","volume":"2024 47","pages":"1-7"},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11872227/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143544839","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Screen Magnification for Readers with Low Vision: A Study on Usability and Performance. 低视力读者的屏幕放大功能:可用性和性能研究
Pub Date : 2023-10-01 Epub Date: 2023-10-22
Meini Tang, Roberto Manduchi, Susana Chung, Raquel Prado

We present a study with 20 participants with low vision who operated two types of screen magnification (lens and full) on a laptop computer to read two types of document (text and web page). Our purposes were to comparatively assess the two magnification modalities, and to obtain some insight into how people with low vision use the mouse to control the center of magnification. These observations may inform the design of systems for the automatic control of the center of magnification. Our results show that there were no significant differences in reading performances or in subjective preferences between the two magnification modes. However, when using the lens mode, our participants adopted more consistent and uniform mouse motion patterns, while longer and more frequent pauses and shorter overall path lengths were measured using the full mode. Analysis of the distribution of gaze points (as measured by a gaze tracker) using the full mode shows that, when reading a text document, most participants preferred to move the area of interest to a specific region of the screen.

我们对 20 名低视能参与者进行了一项研究,他们在笔记本电脑上操作两种屏幕放大方式(镜头放大和完全放大)来阅读两种类型的文档(文本和网页)。我们的目的是对两种放大模式进行比较评估,并深入了解低视力者如何使用鼠标控制放大中心。这些观察结果可为设计自动控制放大中心的系统提供参考。我们的研究结果表明,两种放大模式在阅读效果和主观偏好上没有明显差异。不过,在使用透镜模式时,我们的参与者采用了更一致、更均匀的鼠标运动模式,而在使用完整模式时,他们的停顿时间更长、更频繁,总体路径长度也更短。对使用全屏模式时的注视点分布(由注视跟踪器测量)进行的分析表明,在阅读文本文档时,大多数参与者更倾向于将感兴趣的区域移动到屏幕的特定区域。
{"title":"Screen Magnification for Readers with Low Vision: A Study on Usability and Performance.","authors":"Meini Tang, Roberto Manduchi, Susana Chung, Raquel Prado","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>We present a study with 20 participants with low vision who operated two types of screen magnification (lens and full) on a laptop computer to read two types of document (text and web page). Our purposes were to comparatively assess the two magnification modalities, and to obtain some insight into how people with low vision use the mouse to control the center of magnification. These observations may inform the design of systems for the automatic control of the center of magnification. Our results show that there were no significant differences in reading performances or in subjective preferences between the two magnification modes. However, when using the lens mode, our participants adopted more consistent and uniform mouse motion patterns, while longer and more frequent pauses and shorter overall path lengths were measured using the full mode. Analysis of the distribution of gaze points (as measured by a gaze tracker) using the full mode shows that, when reading a text document, most participants preferred to move the area of interest to a specific region of the screen.</p>","PeriodicalId":72321,"journal":{"name":"ASSETS. Annual ACM Conference on Assistive Technologies","volume":"2023 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10923554/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140095279","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Blind Users Accessing Their Training Images in Teachable Object Recognizers. 盲人用户在可教物体识别器中访问自己的训练图像
Pub Date : 2022-10-01 Epub Date: 2022-10-22 DOI: 10.1145/3517428.3544824
Jonggi Hong, Jaina Gandhi, Ernest Essuah Mensah, Farnaz Zamiri Zeraati, Ebrima Haddy Jarjue, Kyungjun Lee, Hernisa Kacorri

Teachable object recognizers provide a solution for a very practical need for blind people - instance level object recognition. They assume one can visually inspect the photos they provide for training, a critical and inaccessible step for those who are blind. In this work, we engineer data descriptors that address this challenge. They indicate in real time whether the object in the photo is cropped or too small, a hand is included, the photos is blurred, and how much photos vary from each other. Our descriptors are built into open source testbed iOS app, called MYCam. In a remote user study in (N = 12) blind participants' homes, we show how descriptors, even when error-prone, support experimentation and have a positive impact in the quality of training set that can translate to model performance though this gain is not uniform. Participants found the app simple to use indicating that they could effectively train it and that the descriptors were useful. However, many found the training being tedious, opening discussions around the need for balance between information, time, and cognitive load.

可教物体识别器为盲人的一个非常实际的需求提供了解决方案--实例级物体识别。它们假定人们可以目测它们提供的用于训练的照片,而对于盲人来说,这是一个关键且难以接近的步骤。在这项工作中,我们设计了数据描述符来应对这一挑战。它们能实时显示照片中的物体是否被裁剪或太小、是否包含一只手、照片是否模糊以及照片之间的差异程度。我们的描述符内置于名为 MYCam 的开源测试平台 iOS 应用程序中。在盲人参与者家中进行的远程用户研究(N = 12)中,我们展示了描述符(即使容易出错)是如何支持实验并对训练集的质量产生积极影响的,这种影响可以转化为模型性能,尽管这种增益并不一致。参与者认为该应用简单易用,表明他们可以有效地对其进行训练,而且描述符也很有用。不过,许多人认为训练很枯燥乏味,因此开始讨论在信息、时间和认知负荷之间保持平衡的必要性。
{"title":"Blind Users Accessing Their Training Images in Teachable Object Recognizers.","authors":"Jonggi Hong, Jaina Gandhi, Ernest Essuah Mensah, Farnaz Zamiri Zeraati, Ebrima Haddy Jarjue, Kyungjun Lee, Hernisa Kacorri","doi":"10.1145/3517428.3544824","DOIUrl":"10.1145/3517428.3544824","url":null,"abstract":"<p><p>Teachable object recognizers provide a solution for a very practical need for blind people - instance level object recognition. They assume one can visually inspect the photos they provide for training, a critical and inaccessible step for those who are blind. In this work, we engineer data descriptors that address this challenge. They indicate in real time whether the object in the photo is cropped or too small, a hand is included, the photos is blurred, and how much photos vary from each other. Our descriptors are built into open source testbed iOS app, called MYCam. In a remote user study in (<i>N</i> = 12) blind participants' homes, we show how descriptors, even when error-prone, support experimentation and have a positive impact in the quality of training set that can translate to model performance though this gain is not uniform. Participants found the app simple to use indicating that they could effectively train it and that the descriptors were useful. However, many found the training being tedious, opening discussions around the need for balance between information, time, and cognitive load.</p>","PeriodicalId":72321,"journal":{"name":"ASSETS. Annual ACM Conference on Assistive Technologies","volume":"2022 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10008526/pdf/nihms-1869981.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9111608","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mobile Phone Use by People with Mild to Moderate Dementia: Uncovering Challenges and Identifying Opportunities: Mobile Phone Use by People with Mild to Moderate Dementia. 轻至中度痴呆患者的手机使用:发现挑战和识别机会:轻至中度痴呆患者的手机使用。
Pub Date : 2022-10-01 Epub Date: 2022-10-22 DOI: 10.1145/3517428.3544809
Emma Dixon, Rain Michaels, Xiang Xiao, Yu Zhong, Patrick Clary, Ajit Narayanan, Robin Brewer, Amanda Lazar

With the rising usage of mobile phones by people with mild dementia, and the documented barriers to technology use that exist for people with dementia, there is an open opportunity to study the specifics of mobile phone use by people with dementia. In this work we provide a first step towards filling this gap through an interview study with fourteen people with mild to moderate dementia. Our analysis yields insights into mobile phone use by people with mild to moderate dementia, challenges they experience with mobile phone use, and their ideas to address these challenges. Based on these findings, we discuss design opportunities to help achieve more accessible and supportive technology use for people with dementia. Our work opens up new opportunities for the design of systems focused on augmenting and enhancing the abilities of people with dementia.

随着轻度痴呆症患者越来越多地使用移动电话,以及痴呆症患者在使用技术方面存在的记录障碍,有机会研究痴呆症患者使用移动电话的具体情况。在这项工作中,我们通过对14名轻度至中度痴呆症患者的访谈研究,为填补这一空白提供了第一步。我们的分析深入了解了轻度至中度痴呆症患者的手机使用情况,他们在使用手机时遇到的挑战,以及他们应对这些挑战的想法。基于这些发现,我们讨论了设计机会,以帮助痴呆症患者实现更容易获得和支持性的技术使用。我们的工作为设计侧重于增强和提高痴呆症患者能力的系统开辟了新的机会。
{"title":"Mobile Phone Use by People with Mild to Moderate Dementia: Uncovering Challenges and Identifying Opportunities: Mobile Phone Use by People with Mild to Moderate Dementia.","authors":"Emma Dixon, Rain Michaels, Xiang Xiao, Yu Zhong, Patrick Clary, Ajit Narayanan, Robin Brewer, Amanda Lazar","doi":"10.1145/3517428.3544809","DOIUrl":"10.1145/3517428.3544809","url":null,"abstract":"<p><p>With the rising usage of mobile phones by people with mild dementia, and the documented barriers to technology use that exist for people with dementia, there is an open opportunity to study the specifics of mobile phone use by people with dementia. In this work we provide a first step towards filling this gap through an interview study with fourteen people with mild to moderate dementia. Our analysis yields insights into mobile phone use by people with mild to moderate dementia, challenges they experience with mobile phone use, and their ideas to address these challenges. Based on these findings, we discuss design opportunities to help achieve more accessible and supportive technology use for people with dementia. Our work opens up new opportunities for the design of systems focused on augmenting and enhancing the abilities of people with dementia.</p>","PeriodicalId":72321,"journal":{"name":"ASSETS. Annual ACM Conference on Assistive Technologies","volume":"2022 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10202486/pdf/nihms-1865459.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9582599","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Data Representativeness in Accessibility Datasets: A Meta-Analysis. 可访问性数据集的数据代表性:一个元分析。
Pub Date : 2022-10-01 Epub Date: 2022-10-22 DOI: 10.1145/3517428.3544826
Rie Kamikubo, Lining Wang, Crystal Marte, Amnah Mahmood, Hernisa Kacorri

As data-driven systems are increasingly deployed at scale, ethical concerns have arisen around unfair and discriminatory outcomes for historically marginalized groups that are underrepresented in training data. In response, work around AI fairness and inclusion has called for datasets that are representative of various demographic groups. In this paper, we contribute an analysis of the representativeness of age, gender, and race & ethnicity in accessibility datasets-datasets sourced from people with disabilities and older adults-that can potentially play an important role in mitigating bias for inclusive AI-infused applications. We examine the current state of representation within datasets sourced by people with disabilities by reviewing publicly-available information of 190 datasets, we call these accessibility datasets. We find that accessibility datasets represent diverse ages, but have gender and race representation gaps. Additionally, we investigate how the sensitive and complex nature of demographic variables makes classification difficult and inconsistent (e.g., gender, race & ethnicity), with the source of labeling often unknown. By reflecting on the current challenges and opportunities for representation of disabled data contributors, we hope our effort expands the space of possibility for greater inclusion of marginalized communities in AI-infused systems.

随着数据驱动系统越来越多地大规模部署,对于在培训数据中代表性不足的历史边缘化群体,出现了不公平和歧视性结果的伦理担忧。作为回应,围绕人工智能公平性和包容性的工作需要能够代表不同人口群体的数据集。在本文中,我们分析了无障碍数据集中年龄、性别、种族和民族的代表性,这些数据集来自残疾人和老年人,这可能在减轻包容性人工智能应用的偏见方面发挥重要作用。我们通过审查190个数据集(我们称之为无障碍数据集)的公开信息,检查了残疾人数据集中的当前表现状态。我们发现可访问性数据集代表不同的年龄,但存在性别和种族代表性差距。此外,我们研究了人口统计变量的敏感性和复杂性如何使分类困难和不一致(例如,性别,种族和民族),标签的来源通常未知。通过反思当前残疾数据贡献者所面临的挑战和机遇,我们希望我们的努力能够扩大将边缘化社区更大程度地纳入人工智能系统的可能性空间。
{"title":"Data Representativeness in Accessibility Datasets: A Meta-Analysis.","authors":"Rie Kamikubo, Lining Wang, Crystal Marte, Amnah Mahmood, Hernisa Kacorri","doi":"10.1145/3517428.3544826","DOIUrl":"10.1145/3517428.3544826","url":null,"abstract":"<p><p>As data-driven systems are increasingly deployed at scale, ethical concerns have arisen around unfair and discriminatory outcomes for historically marginalized groups that are underrepresented in training data. In response, work around AI fairness and inclusion has called for datasets that are representative of various demographic groups. In this paper, we contribute an analysis of the representativeness of age, gender, and race & ethnicity in accessibility datasets-datasets sourced from people with disabilities and older adults-that can potentially play an important role in mitigating bias for inclusive AI-infused applications. We examine the current state of representation within datasets sourced by people with disabilities by reviewing publicly-available information of 190 datasets, we call these accessibility datasets. We find that accessibility datasets represent diverse ages, but have gender and race representation gaps. Additionally, we investigate how the sensitive and complex nature of demographic variables makes classification difficult and inconsistent (<i>e.g.</i>, gender, race & ethnicity), with the source of labeling often unknown. By reflecting on the current challenges and opportunities for representation of disabled data contributors, we hope our effort expands the space of possibility for greater inclusion of marginalized communities in AI-infused systems.</p>","PeriodicalId":72321,"journal":{"name":"ASSETS. Annual ACM Conference on Assistive Technologies","volume":"2022 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10024595/pdf/nihms-1869788.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9153813","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
ASSETS. Annual ACM Conference on Assistive Technologies
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1