首页 > 最新文献

Proceedings. 20th International Conference on Data Engineering最新文献

英文 中文
A type-safe object-oriented solution for the dynamic construction of queries 用于查询动态构造的类型安全的面向对象解决方案
Pub Date : 2004-03-30 DOI: 10.1109/ICDE.2004.1320073
Peter Rosenthal
Many object-oriented applications use large numbers of structurally different database queries. With current technology, writing applications that generate queries at runtime is difficult and error-prone. FROQUE, a framework for object-oriented queries, provides a secure and purely object-oriented solution to access relational databases. As such, it is easy to use for object-oriented programmers and with the help of object-oriented compilers it guarantees that queries formulated in the object-oriented world at execution time result in correct SQL queries. Thus, FROQUE is an improvement over existing database frameworks such as Apache OJB, the object relational bridge, which are not strongly typed and can lead to runtime errors.
许多面向对象的应用程序使用大量结构不同的数据库查询。使用当前的技术,编写在运行时生成查询的应用程序很困难,而且容易出错。面向对象查询的框架FROQUE提供了访问关系数据库的安全和纯粹面向对象的解决方案。因此,对于面向对象的程序员来说,它很容易使用,并且在面向对象编译器的帮助下,它保证在执行时在面向对象世界中制定的查询会产生正确的SQL查询。因此,FROQUE是对现有数据库框架(如Apache OJB,对象关系桥)的改进,后者不是强类型的,可能导致运行时错误。
{"title":"A type-safe object-oriented solution for the dynamic construction of queries","authors":"Peter Rosenthal","doi":"10.1109/ICDE.2004.1320073","DOIUrl":"https://doi.org/10.1109/ICDE.2004.1320073","url":null,"abstract":"Many object-oriented applications use large numbers of structurally different database queries. With current technology, writing applications that generate queries at runtime is difficult and error-prone. FROQUE, a framework for object-oriented queries, provides a secure and purely object-oriented solution to access relational databases. As such, it is easy to use for object-oriented programmers and with the help of object-oriented compilers it guarantees that queries formulated in the object-oriented world at execution time result in correct SQL queries. Thus, FROQUE is an improvement over existing database frameworks such as Apache OJB, the object relational bridge, which are not strongly typed and can lead to runtime errors.","PeriodicalId":358862,"journal":{"name":"Proceedings. 20th International Conference on Data Engineering","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121653131","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bitmap-tree indexing for set operations on free text 位图树索引在自由文本上的集合操作
Pub Date : 2004-03-30 DOI: 10.1109/ICDE.2004.1320067
Ilias Nitsos, Georgios Evangelidis, D. Dervos
Here we report on our implementation of a hybrid-indexing scheme (bitmap-tree) that combines the advantages of bitmap indexing and file inversion. The results we obtained are compared to those of the compressed inverted file index. Both storage overhead and query processing efficiency are taken into consideration. The proposed new method is shown to excel in handling queries involving set operations. For general-purpose user queries, the bitmap-tree is shown to perform as good as the compressed inverted file index.
在这里,我们报告了一种混合索引方案(位图树)的实现,它结合了位图索引和文件反转的优点。将所得结果与压缩后的倒排文件索引结果进行了比较。同时考虑了存储开销和查询处理效率。所提出的新方法在处理涉及集合操作的查询方面表现优异。对于一般用途的用户查询,位图树的表现与压缩的倒排文件索引一样好。
{"title":"Bitmap-tree indexing for set operations on free text","authors":"Ilias Nitsos, Georgios Evangelidis, D. Dervos","doi":"10.1109/ICDE.2004.1320067","DOIUrl":"https://doi.org/10.1109/ICDE.2004.1320067","url":null,"abstract":"Here we report on our implementation of a hybrid-indexing scheme (bitmap-tree) that combines the advantages of bitmap indexing and file inversion. The results we obtained are compared to those of the compressed inverted file index. Both storage overhead and query processing efficiency are taken into consideration. The proposed new method is shown to excel in handling queries involving set operations. For general-purpose user queries, the bitmap-tree is shown to perform as good as the compressed inverted file index.","PeriodicalId":358862,"journal":{"name":"Proceedings. 20th International Conference on Data Engineering","volume":"73 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132218395","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Approximate selection queries over imprecise data 对不精确数据的近似选择查询
Pub Date : 2004-03-30 DOI: 10.1109/ICDE.2004.1319991
Iosif Lazaridis, S. Mehrotra
We examine the problem of evaluating selection queries over imprecisely represented objects. Such objects are used either because they are much smaller in size than the precise ones (e.g., compressed versions of time series), or as imprecise replicas of fast-changing objects across the network (e.g., interval approximations for time-varying sensor readings). It may be impossible to determine whether an imprecise object meets the selection predicate. Additionally, the objects appearing in the output are also imprecise. Retrieving the precise objects themselves (at additional cost) can be used to increase the quality of the reported answer. We allow queries to specify their own answer quality requirements. We show how the query evaluation system may do the minimal amount of work to meet these requirements. Our work presents two important contributions: first, by considering queries with set-based answers, rather than the approximate aggregate queries over numerical data examined in the literature; second, by aiming to minimize the combined cost of both data processing and probe operations in a single framework. Thus, we establish that the answer accuracy/performance tradeoff can be realized in a more general setting than previously seen.
我们研究在不精确表示的对象上评估选择查询的问题。使用这些对象,要么是因为它们的尺寸比精确对象(例如,时间序列的压缩版本)小得多,要么是因为它们是网络中快速变化对象的不精确复制品(例如,时变传感器读数的间隔近似)。可能无法确定不精确的对象是否满足选择谓词。此外,输出中出现的对象也是不精确的。检索精确的对象本身(需要额外的成本)可以用来提高报告答案的质量。我们允许查询指定它们自己的回答质量要求。我们将展示查询评估系统如何以最少的工作量来满足这些需求。我们的工作提出了两个重要的贡献:首先,通过考虑具有基于集合的答案的查询,而不是在文献中检查的数值数据的近似聚合查询;其次,通过在单个框架中最小化数据处理和探测操作的综合成本。因此,我们确定答案准确性/性能权衡可以在比以前看到的更一般的设置中实现。
{"title":"Approximate selection queries over imprecise data","authors":"Iosif Lazaridis, S. Mehrotra","doi":"10.1109/ICDE.2004.1319991","DOIUrl":"https://doi.org/10.1109/ICDE.2004.1319991","url":null,"abstract":"We examine the problem of evaluating selection queries over imprecisely represented objects. Such objects are used either because they are much smaller in size than the precise ones (e.g., compressed versions of time series), or as imprecise replicas of fast-changing objects across the network (e.g., interval approximations for time-varying sensor readings). It may be impossible to determine whether an imprecise object meets the selection predicate. Additionally, the objects appearing in the output are also imprecise. Retrieving the precise objects themselves (at additional cost) can be used to increase the quality of the reported answer. We allow queries to specify their own answer quality requirements. We show how the query evaluation system may do the minimal amount of work to meet these requirements. Our work presents two important contributions: first, by considering queries with set-based answers, rather than the approximate aggregate queries over numerical data examined in the literature; second, by aiming to minimize the combined cost of both data processing and probe operations in a single framework. Thus, we establish that the answer accuracy/performance tradeoff can be realized in a more general setting than previously seen.","PeriodicalId":358862,"journal":{"name":"Proceedings. 20th International Conference on Data Engineering","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129171172","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
Probe, cluster, and discover: focused extraction of QA-Pagelets from the deep Web 探测、聚类和发现:从深度网络集中提取QA-Pagelets
Pub Date : 2004-03-30 DOI: 10.1109/ICDE.2004.1319988
James Caverlee, Ling Liu, David J. Buttler
We introduce the concept of a QA-Pagelet to refer to the content region in a dynamic page that contains query matches. We present THOR, a scalable and efficient mining system for discovering and extracting QA-Pagelets from the deep Web. A unique feature of THOR is its two-phase extraction framework. In the first phase, pages from a deep Web site are grouped into distinct clusters of structurally-similar pages. In the second phase, pages from each page cluster are examined through a subtree filtering algorithm that exploits the structural and content similarity at subtree level to identify the QA-Pagelets.
我们引入QA-Pagelet的概念来引用包含查询匹配的动态页面中的内容区域。我们提出了THOR,一个可扩展且高效的挖掘系统,用于从深度网络中发现和提取QA-Pagelets。THOR的一个独特之处在于它的两相提取框架。在第一阶段,来自深度Web站点的页面被分组到结构相似的不同页面集群中。在第二阶段,通过子树过滤算法检查每个页面簇中的页面,该算法利用子树级别的结构和内容相似性来识别QA-Pagelets。
{"title":"Probe, cluster, and discover: focused extraction of QA-Pagelets from the deep Web","authors":"James Caverlee, Ling Liu, David J. Buttler","doi":"10.1109/ICDE.2004.1319988","DOIUrl":"https://doi.org/10.1109/ICDE.2004.1319988","url":null,"abstract":"We introduce the concept of a QA-Pagelet to refer to the content region in a dynamic page that contains query matches. We present THOR, a scalable and efficient mining system for discovering and extracting QA-Pagelets from the deep Web. A unique feature of THOR is its two-phase extraction framework. In the first phase, pages from a deep Web site are grouped into distinct clusters of structurally-similar pages. In the second phase, pages from each page cluster are examined through a subtree filtering algorithm that exploits the structural and content similarity at subtree level to identify the QA-Pagelets.","PeriodicalId":358862,"journal":{"name":"Proceedings. 20th International Conference on Data Engineering","volume":"84 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131813972","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 48
A Web-services architecture for efficient XML data exchange 用于高效XML数据交换的web服务体系结构
Pub Date : 2004-03-30 DOI: 10.1109/ICDE.2004.1320024
S. Amer-Yahia, Y. Kotidis
Business applications often exchange large amounts of enterprise data stored in legacy systems. The advent of XML as a standard specification format has improved applications interoperability. However, optimizing the performance of XML data exchange, in particular, when data volumes are large, is still in its infancy. Quite often, the target system has to undo some of the work the source did to assemble documents in order to map XML elements into its own data structures. This publish&map process is both resource and time consuming. In this paper, we develop a middle-tier Web services architecture to optimize the exchange of large XML data volumes. The key idea is to allow systems to negotiate the data exchange process using an extension to WSDL. The source (target) can specify document fragments that it is willing to produce (consume). Given these fragmentations, the middleware instruments the data exchange process between the two systems to minimize the number of necessary operations and optimize the distributed processing between the source and the target systems. We show that our new exchange paradigm outperforms publish&map and enables more flexible scenarios without necessitating substantial modifications to the underlying systems.
业务应用程序经常交换存储在遗留系统中的大量企业数据。XML作为标准规范格式的出现改进了应用程序的互操作性。但是,优化XML数据交换的性能,特别是在数据量很大的情况下,仍然处于起步阶段。为了将XML元素映射到自己的数据结构中,目标系统经常必须撤销源系统为组装文档所做的一些工作。这个发布和映射过程既耗费资源又耗费时间。在本文中,我们开发了一个中间层Web服务体系结构来优化大型XML数据量的交换。关键思想是允许系统使用WSDL扩展来协商数据交换过程。源(目标)可以指定它愿意生成(使用)的文档片段。给定这些片段,中间件将在两个系统之间配置数据交换过程,以最小化必要操作的数量,并优化源系统和目标系统之间的分布式处理。我们展示了我们的新交换范例优于publish&map,并且在不需要对底层系统进行实质性修改的情况下支持更灵活的场景。
{"title":"A Web-services architecture for efficient XML data exchange","authors":"S. Amer-Yahia, Y. Kotidis","doi":"10.1109/ICDE.2004.1320024","DOIUrl":"https://doi.org/10.1109/ICDE.2004.1320024","url":null,"abstract":"Business applications often exchange large amounts of enterprise data stored in legacy systems. The advent of XML as a standard specification format has improved applications interoperability. However, optimizing the performance of XML data exchange, in particular, when data volumes are large, is still in its infancy. Quite often, the target system has to undo some of the work the source did to assemble documents in order to map XML elements into its own data structures. This publish&map process is both resource and time consuming. In this paper, we develop a middle-tier Web services architecture to optimize the exchange of large XML data volumes. The key idea is to allow systems to negotiate the data exchange process using an extension to WSDL. The source (target) can specify document fragments that it is willing to produce (consume). Given these fragmentations, the middleware instruments the data exchange process between the two systems to minimize the number of necessary operations and optimize the distributed processing between the source and the target systems. We show that our new exchange paradigm outperforms publish&map and enables more flexible scenarios without necessitating substantial modifications to the underlying systems.","PeriodicalId":358862,"journal":{"name":"Proceedings. 20th International Conference on Data Engineering","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131584361","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 38
Efficient incremental validation of XML documents 对XML文档进行有效的增量验证
Pub Date : 2004-03-30 DOI: 10.1109/ICDE.2004.1320036
Denilson Barbosa, A. Mendelzon, L. Libkin, L. Mignet, M. Arenas
We discuss incremental validation of XML documents with respect to DTDs and XML schema definitions. We consider insertions and deletions of subtrees, as opposed to leaf nodes only, and we also consider the validation of ID and IDREF attributes. For arbitrary schemas, we give a worst-case n log n time and linear space algorithm, and show that it often is far superior to revalidation from scratch. We present two classes of schemas, which capture most real-life DTDs, and show that they admit a logarithmic time incremental validation algorithm that, in many cases, requires only constant auxiliary space. We then discuss an implementation of these algorithms that is independent of, and can be customized for different storage mechanisms for XML. Finally, we present extensive experimental results showing that our approach is highly efficient and scalable.
我们将根据dtd和XML模式定义讨论XML文档的增量验证。我们考虑子树的插入和删除,而不是只考虑叶节点,我们还考虑ID和IDREF属性的验证。对于任意模式,我们给出了最坏情况下的n log n时间和线性空间算法,并表明它通常比从头开始重新验证要好得多。我们提出了两类模式,它们捕获了大多数现实生活中的dtd,并表明它们支持对数时间增量验证算法,在许多情况下,该算法只需要恒定的辅助空间。然后我们讨论这些算法的实现,这些算法独立于XML的不同存储机制,并且可以针对不同的存储机制进行定制。最后,我们给出了大量的实验结果,表明我们的方法是高效和可扩展的。
{"title":"Efficient incremental validation of XML documents","authors":"Denilson Barbosa, A. Mendelzon, L. Libkin, L. Mignet, M. Arenas","doi":"10.1109/ICDE.2004.1320036","DOIUrl":"https://doi.org/10.1109/ICDE.2004.1320036","url":null,"abstract":"We discuss incremental validation of XML documents with respect to DTDs and XML schema definitions. We consider insertions and deletions of subtrees, as opposed to leaf nodes only, and we also consider the validation of ID and IDREF attributes. For arbitrary schemas, we give a worst-case n log n time and linear space algorithm, and show that it often is far superior to revalidation from scratch. We present two classes of schemas, which capture most real-life DTDs, and show that they admit a logarithmic time incremental validation algorithm that, in many cases, requires only constant auxiliary space. We then discuss an implementation of these algorithms that is independent of, and can be customized for different storage mechanisms for XML. Finally, we present extensive experimental results showing that our approach is highly efficient and scalable.","PeriodicalId":358862,"journal":{"name":"Proceedings. 20th International Conference on Data Engineering","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115739248","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 78
BOSS: browsing OPTICS-plots for similarity search 老板:浏览光学图进行相似度搜索
Pub Date : 2004-03-30 DOI: 10.1109/ICDE.2004.1320086
S. Brecheisen, H. Kriegel, Peer Kröger, M. Pfeifle, Maximilian Viermetz, Marco Pötke
An increasing number of database applications have emerged for which efficient and effective support for similarity search is substantial. Particularly, the task of finding similar shapes in 2D and 3D becomes more and more important. Examples for new applications that require the retrieval of similar 3D objects include databases for molecular biology, medical imaging and computer aided design. Hierarchical clustering was shown to be effective for evaluating similarity models. Furthermore, visually analyzing cluster hierarchies helps the user, e.g. an engineer, to find and group similar objects. We present an interactive browsing tool called BOSS (browsing OPTICS-plots for similarity search), which utilizes solid automatic cluster recognition and extraction of meaningful cluster representatives in order to provide the user with significant and quick information.
越来越多的数据库应用程序已经出现,对相似度搜索的高效和有效支持是重要的。特别是,在二维和三维中寻找相似形状的任务变得越来越重要。需要检索类似3D对象的新应用示例包括分子生物学、医学成像和计算机辅助设计的数据库。层次聚类被证明是评估相似模型的有效方法。此外,可视化地分析聚类层次结构可以帮助用户(例如工程师)找到并分组相似的对象。我们提出了一种交互式浏览工具BOSS(浏览OPTICS-plots for similarity search),它利用实体自动聚类识别和提取有意义的聚类代表来为用户提供有意义和快速的信息。
{"title":"BOSS: browsing OPTICS-plots for similarity search","authors":"S. Brecheisen, H. Kriegel, Peer Kröger, M. Pfeifle, Maximilian Viermetz, Marco Pötke","doi":"10.1109/ICDE.2004.1320086","DOIUrl":"https://doi.org/10.1109/ICDE.2004.1320086","url":null,"abstract":"An increasing number of database applications have emerged for which efficient and effective support for similarity search is substantial. Particularly, the task of finding similar shapes in 2D and 3D becomes more and more important. Examples for new applications that require the retrieval of similar 3D objects include databases for molecular biology, medical imaging and computer aided design. Hierarchical clustering was shown to be effective for evaluating similarity models. Furthermore, visually analyzing cluster hierarchies helps the user, e.g. an engineer, to find and group similar objects. We present an interactive browsing tool called BOSS (browsing OPTICS-plots for similarity search), which utilizes solid automatic cluster recognition and extraction of meaningful cluster representatives in order to provide the user with significant and quick information.","PeriodicalId":358862,"journal":{"name":"Proceedings. 20th International Conference on Data Engineering","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121785110","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
FLYINGDOC: an architecture for distributed, user-friendly, and personalized information systems FLYINGDOC:分布式、用户友好和个性化信息系统的架构
Pub Date : 2004-03-30 DOI: 10.1109/ICDE.2004.1320078
I. Bruder, A. Zeitz, Holger Meyer, B. Hänsel, A. Heuer
The need for personal information management using distributed, user-friendly, and personalized document management systems is obvious. State of the art document management systems such as digital libraries provide support for the whole document lifecycle. To enhance such document management systems to get a personalized, distributed and user-friendly information system we present techniques for a simple import of collections, documents, and data, for generic and concrete data modeling, replication, and, personalization. These techniques were employed for the implementation of a personal conference assistant, which was used for the first time at the VLDB conference 2003 in Berlin, Germany. Our client-server architecture provides an information server with different services and different kinds of clients. These services comprise a distribution and replication service, a collection integration service, a data management unit, and, a query processing service.
使用分布式的、用户友好的和个性化的文档管理系统进行个人信息管理的需求是显而易见的。最先进的文档管理系统,如数字图书馆,为整个文档生命周期提供支持。为了增强这样的文档管理系统,使其成为个性化、分布式和用户友好的信息系统,我们提出了用于简单导入集合、文档和数据的技术,用于通用和具体的数据建模、复制和个性化。这些技术被用于实现个人会议助理,并在2003年德国柏林举行的VLDB会议上首次使用。我们的客户机-服务器体系结构提供了具有不同服务和不同类型客户机的信息服务器。这些服务包括分发和复制服务、集合集成服务、数据管理单元和查询处理服务。
{"title":"FLYINGDOC: an architecture for distributed, user-friendly, and personalized information systems","authors":"I. Bruder, A. Zeitz, Holger Meyer, B. Hänsel, A. Heuer","doi":"10.1109/ICDE.2004.1320078","DOIUrl":"https://doi.org/10.1109/ICDE.2004.1320078","url":null,"abstract":"The need for personal information management using distributed, user-friendly, and personalized document management systems is obvious. State of the art document management systems such as digital libraries provide support for the whole document lifecycle. To enhance such document management systems to get a personalized, distributed and user-friendly information system we present techniques for a simple import of collections, documents, and data, for generic and concrete data modeling, replication, and, personalization. These techniques were employed for the implementation of a personal conference assistant, which was used for the first time at the VLDB conference 2003 in Berlin, Germany. Our client-server architecture provides an information server with different services and different kinds of clients. These services comprise a distribution and replication service, a collection integration service, a data management unit, and, a query processing service.","PeriodicalId":358862,"journal":{"name":"Proceedings. 20th International Conference on Data Engineering","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116933605","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
SQLCM: a continuous monitoring framework for relational database engines SQLCM:关系数据库引擎的连续监视框架
Pub Date : 2004-03-30 DOI: 10.1109/ICDE.2004.1320020
S. Chaudhuri, A. König, Vivek R. Narasayya
The ability to monitor a database server is crucial for effective database administration. Today's commercial database systems support two basic mechanisms for monitoring: (a) obtaining a snapshot of counters to capture current state, and (b) logging events in the server to a table/file to capture history. We show that for a large class of important database administration tasks the above mechanisms are inadequate in functionality or performance. We present an infrastructure called SQLCM that enables continuous monitoring inside the database server and that has the ability to automatically take actions based on monitoring. We describe the implementation of SQLCM in Microsoft SQL Server and show how several common and important monitoring tasks can be easily specified in SQLCM. Our experimental evaluation indicates that SQLCM imposes low overhead on normal server execution end enables monitoring tasks on a production server that would be too expensive using today's monitoring mechanisms.
监视数据库服务器的能力对于有效的数据库管理至关重要。今天的商业数据库系统支持两种基本的监视机制:(a)获取计数器的快照以捕获当前状态;(b)将服务器中的事件记录到表/文件中以捕获历史。我们表明,对于大量重要的数据库管理任务,上述机制在功能或性能上都是不够的。我们提出了一种名为SQLCM的基础设施,它支持在数据库服务器内部进行连续监视,并且能够根据监视自动采取操作。我们将描述SQLCM在Microsoft SQL Server中的实现,并展示如何在SQLCM中轻松指定几个常见且重要的监视任务。我们的实验评估表明,SQLCM对普通服务器执行端的开销很低,因此可以在生产服务器上监视任务,而使用目前的监视机制会造成太大的开销。
{"title":"SQLCM: a continuous monitoring framework for relational database engines","authors":"S. Chaudhuri, A. König, Vivek R. Narasayya","doi":"10.1109/ICDE.2004.1320020","DOIUrl":"https://doi.org/10.1109/ICDE.2004.1320020","url":null,"abstract":"The ability to monitor a database server is crucial for effective database administration. Today's commercial database systems support two basic mechanisms for monitoring: (a) obtaining a snapshot of counters to capture current state, and (b) logging events in the server to a table/file to capture history. We show that for a large class of important database administration tasks the above mechanisms are inadequate in functionality or performance. We present an infrastructure called SQLCM that enables continuous monitoring inside the database server and that has the ability to automatically take actions based on monitoring. We describe the implementation of SQLCM in Microsoft SQL Server and show how several common and important monitoring tasks can be easily specified in SQLCM. Our experimental evaluation indicates that SQLCM imposes low overhead on normal server execution end enables monitoring tasks on a production server that would be too expensive using today's monitoring mechanisms.","PeriodicalId":358862,"journal":{"name":"Proceedings. 20th International Conference on Data Engineering","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117069682","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 32
Proving ownership over categorical data 证明对分类数据的所有权
Pub Date : 2004-03-30 DOI: 10.1109/ICDE.2004.1320029
R. Sion
This paper introduces a novel method of rights protection for categorical data through watermarking. We discover new watermark embedding channels for relational data with categorical types. We design novel watermark encoding algorithms and analyze important theoretical bounds including mark vulnerability. While fully preserving data quality requirements, our solution survives important attacks, such as subset selection and random alterations. Mark detection is fully "blind" in that it doesn't require the original data, an important characteristic especially in the case of massive data. We propose various improvements and alternative encoding methods. We perform validation experiments by watermarking the outsourced Wal-Mart sales data available at our institute. We prove (experimentally and by analysis) our solution to be extremely resilient to both alteration and data loss attacks, for example tolerating up to 80% data loss with a watermark alteration of only 25%.
介绍了一种利用水印对分类数据进行版权保护的新方法。我们为具有分类类型的关系数据发现了新的水印嵌入通道。设计了新的水印编码算法,分析了水印脆弱性等重要理论边界。在完全保留数据质量要求的同时,我们的解决方案能够承受重要的攻击,例如子集选择和随机更改。标记检测是完全“盲”的,因为它不需要原始数据,这是一个重要的特征,特别是在海量数据的情况下。我们提出了各种改进和替代编码方法。我们通过对我们研究所提供的外包沃尔玛销售数据进行水印来进行验证实验。我们(通过实验和分析)证明我们的解决方案对更改和数据丢失攻击具有极高的弹性,例如,容忍高达80%的数据丢失,水印更改仅为25%。
{"title":"Proving ownership over categorical data","authors":"R. Sion","doi":"10.1109/ICDE.2004.1320029","DOIUrl":"https://doi.org/10.1109/ICDE.2004.1320029","url":null,"abstract":"This paper introduces a novel method of rights protection for categorical data through watermarking. We discover new watermark embedding channels for relational data with categorical types. We design novel watermark encoding algorithms and analyze important theoretical bounds including mark vulnerability. While fully preserving data quality requirements, our solution survives important attacks, such as subset selection and random alterations. Mark detection is fully \"blind\" in that it doesn't require the original data, an important characteristic especially in the case of massive data. We propose various improvements and alternative encoding methods. We perform validation experiments by watermarking the outsourced Wal-Mart sales data available at our institute. We prove (experimentally and by analysis) our solution to be extremely resilient to both alteration and data loss attacks, for example tolerating up to 80% data loss with a watermark alteration of only 25%.","PeriodicalId":358862,"journal":{"name":"Proceedings. 20th International Conference on Data Engineering","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125917393","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 97
期刊
Proceedings. 20th International Conference on Data Engineering
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1