Linking unstructured evidence to structured observations

IF 1.8 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Information Visualization Pub Date : 2021-01-01 DOI:10.1177/1473871620986249
Manuela Waldner, Thomas Geymayer, D. Schmalstieg, M. Sedlmair
{"title":"Linking unstructured evidence to structured observations","authors":"Manuela Waldner, Thomas Geymayer, D. Schmalstieg, M. Sedlmair","doi":"10.1177/1473871620986249","DOIUrl":null,"url":null,"abstract":"Many professionals, like journalists, writers, or consultants, need to acquire information from various sources, make sense of this unstructured evidence, structure their observations, and finally create and deliver their product, such as a report or a presentation. In formative interviews, we found that tools allowing structuring of observations are often disconnected from the corresponding evidence. Therefore, we designed a sensemaking environment with a flexible observation graph that visually ties together evidence in unstructured documents with the user’s structured knowledge. This is achieved through bi-directional deep links between highlighted document portions and nodes in the observation graph. In a controlled study, we compared users’ sensemaking strategies using either the observation graph or a simple text editor on a large display. Results show that the observation graph represents a holistic, compact representation of users’ observations, which can be linked to unstructured evidence on demand. In contrast, users taking textual notes required much more display space to spatially organize source documents containing unstructured evidence. This implies that spatial organization is a powerful strategy to structure observations even if the available space is limited.","PeriodicalId":50360,"journal":{"name":"Information Visualization","volume":"20 1","pages":"47 - 65"},"PeriodicalIF":1.8000,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1177/1473871620986249","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Visualization","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1177/1473871620986249","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, SOFTWARE ENGINEERING","Score":null,"Total":0}
引用次数: 2

Abstract

Many professionals, like journalists, writers, or consultants, need to acquire information from various sources, make sense of this unstructured evidence, structure their observations, and finally create and deliver their product, such as a report or a presentation. In formative interviews, we found that tools allowing structuring of observations are often disconnected from the corresponding evidence. Therefore, we designed a sensemaking environment with a flexible observation graph that visually ties together evidence in unstructured documents with the user’s structured knowledge. This is achieved through bi-directional deep links between highlighted document portions and nodes in the observation graph. In a controlled study, we compared users’ sensemaking strategies using either the observation graph or a simple text editor on a large display. Results show that the observation graph represents a holistic, compact representation of users’ observations, which can be linked to unstructured evidence on demand. In contrast, users taking textual notes required much more display space to spatially organize source documents containing unstructured evidence. This implies that spatial organization is a powerful strategy to structure observations even if the available space is limited.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
将非结构化证据与结构化观察联系起来
许多专业人士,如记者、作家或顾问,需要从各种来源获取信息,理解这些非结构化的证据,构建他们的观察结果,并最终创建和交付他们的产品,如报告或演示。在形成性访谈中,我们发现允许构建观察结果的工具往往与相应的证据脱节。因此,我们设计了一个具有灵活观察图的感知环境,该观察图在视觉上将非结构化文档中的证据与用户的结构化知识联系在一起。这是通过高亮显示的文档部分和观察图中的节点之间的双向深度链接实现的。在一项对照研究中,我们在大屏幕上使用观察图或简单的文本编辑器比较了用户的感知策略。结果表明,观察图代表了用户观察的整体、紧凑的表示,可以根据需要将其与非结构化证据联系起来。相比之下,做文本笔记的用户需要更多的显示空间来空间组织包含非结构化证据的源文档。这意味着,即使可用空间有限,空间组织也是构建观测的有力策略。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Information Visualization
Information Visualization COMPUTER SCIENCE, SOFTWARE ENGINEERING-
CiteScore
5.40
自引率
0.00%
发文量
16
审稿时长
>12 weeks
期刊介绍: Information Visualization is essential reading for researchers and practitioners of information visualization and is of interest to computer scientists and data analysts working on related specialisms. This journal is an international, peer-reviewed journal publishing articles on fundamental research and applications of information visualization. The journal acts as a dedicated forum for the theories, methodologies, techniques and evaluations of information visualization and its applications. The journal is a core vehicle for developing a generic research agenda for the field by identifying and developing the unique and significant aspects of information visualization. Emphasis is placed on interdisciplinary material and on the close connection between theory and practice. This journal is a member of the Committee on Publication Ethics (COPE).
期刊最新文献
Multidimensional data visualization and synchronization for revealing hidden pandemic information Interactive visual formula composition of multidimensional data classifiers Exploring annotation taxonomy in grouped bar charts: A qualitative classroom study Designing complex network visualisations using the wayfinding map metaphor Graph & Network Visualization and Beyond
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1