Preattentive reading and selective attention for document image analysis

C. Faure
{"title":"Preattentive reading and selective attention for document image analysis","authors":"C. Faure","doi":"10.1109/ICDAR.1999.791853","DOIUrl":null,"url":null,"abstract":"PixED (from Pixel to Electronic Document) is aimed at converting document images into structured electronic documents which can be read by a machine for information retrieval. The approach is based on the combination of perception and symbol reading which are the two processes involved when humans detect the organisation of a document. \"Preattentive reading\" denotes the physical segmentation related to perceptual organisation. \"Selective attention\" means that symbol reading is limited to specific sequences of symbols or to pre-attentively selected locations. An OCR provides the primary structured description of the document. PixED improves the quality of this description, completes the physical segmentation and adds a logical description. A distributed software architecture and an incremental strategy are defined to enable the integration of perception and symbol reading. The approach is tested on a set of documents composed of several pages which are gathered from proceedings of scientific conferences.","PeriodicalId":130039,"journal":{"name":"Proceedings of the Fifth International Conference on Document Analysis and Recognition. ICDAR '99 (Cat. No.PR00318)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1999-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the Fifth International Conference on Document Analysis and Recognition. ICDAR '99 (Cat. No.PR00318)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDAR.1999.791853","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

PixED (from Pixel to Electronic Document) is aimed at converting document images into structured electronic documents which can be read by a machine for information retrieval. The approach is based on the combination of perception and symbol reading which are the two processes involved when humans detect the organisation of a document. "Preattentive reading" denotes the physical segmentation related to perceptual organisation. "Selective attention" means that symbol reading is limited to specific sequences of symbols or to pre-attentively selected locations. An OCR provides the primary structured description of the document. PixED improves the quality of this description, completes the physical segmentation and adds a logical description. A distributed software architecture and an incremental strategy are defined to enable the integration of perception and symbol reading. The approach is tested on a set of documents composed of several pages which are gathered from proceedings of scientific conferences.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
文献图像分析的预注意阅读与选择性注意
PixED(从像素到电子文档)的目的是将文档图像转换成结构化的电子文档,使机器能够读取这些电子文档以进行信息检索。该方法是基于感知和符号阅读的结合,这是人类检测文档组织时涉及的两个过程。“前注意阅读”表示与知觉组织相关的物理分割。“选择性注意”意味着符号阅读仅限于特定的符号序列或预先注意选择的位置。OCR提供文档的主要结构化描述。PixED提高了描述的质量,完成了物理分割并添加了逻辑描述。定义了分布式软件架构和增量策略,实现了感知和符号读取的集成。该方法在一组由几页组成的文件上进行了测试,这些文件是从科学会议的会议记录中收集的。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
A two-step algorithm and its parallelization for the generation of minimum containing rectangles for document image segmentation Error correction of Korean courtesy amounts in bank slips using rule information and cross-referencing An off-line signature verification system using an extracted displacement function Quality evaluation of document segmentation results Handwritten numeral string recognition with stroke grouping
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1