首页 > 最新文献

21st International Conference on Data Engineering (ICDE'05)最新文献

英文 中文
A probabilistic XML approach to data integration 数据集成的概率XML方法
Pub Date : 2005-04-05 DOI: 10.1109/ICDE.2005.11
M. V. Keulen, A. D. Keijzer, W. Alink
In mobile and ambient environments, devices need to become autonomous, managing and resolving problems without interference from a user. The database of a (mobile) device can be seen as its knowledge about objects in the 'real world'. Data exchange between small and/or large computing devices can be used to supplement and update this knowledge whenever a connection gets established. In many situations, however, data from different data sources referring to the same real world objects, may conflict. It is the task of the data management system of the device to resolve such conflicts without interference from a user. In this paper, we take a first step in the development of a probabilistic XML DBMS. The main idea is to drop the assumption that data in the database should be certain: subtrees in XML documents may denote possible views on the real world. We formally define the notion of probabilistic XML tree and several operations thereon. We also present an approach for determining a logical semantics for queries on probabilistic XML data. Finally, we introduce an approach for XML data integration where conflicts are resolved by the introduction of possibilities in the database.
在移动和环境环境中,设备需要变得自主,在没有用户干扰的情况下管理和解决问题。(移动)设备的数据库可以被看作是它关于“现实世界”中物体的知识。小型和/或大型计算设备之间的数据交换可用于在建立连接时补充和更新此知识。但是,在许多情况下,来自不同数据源的数据引用相同的现实世界对象可能会发生冲突。在不受用户干扰的情况下解决这些冲突是设备数据管理系统的任务。在本文中,我们迈出了开发概率XML DBMS的第一步。其主要思想是放弃数据库中的数据应该是确定的假设:XML文档中的子树可能表示现实世界中的可能视图。我们正式定义了概率XML树的概念及其上的一些操作。我们还提供了一种确定概率XML数据查询的逻辑语义的方法。最后,我们介绍了一种XML数据集成方法,其中通过在数据库中引入可能性来解决冲突。
{"title":"A probabilistic XML approach to data integration","authors":"M. V. Keulen, A. D. Keijzer, W. Alink","doi":"10.1109/ICDE.2005.11","DOIUrl":"https://doi.org/10.1109/ICDE.2005.11","url":null,"abstract":"In mobile and ambient environments, devices need to become autonomous, managing and resolving problems without interference from a user. The database of a (mobile) device can be seen as its knowledge about objects in the 'real world'. Data exchange between small and/or large computing devices can be used to supplement and update this knowledge whenever a connection gets established. In many situations, however, data from different data sources referring to the same real world objects, may conflict. It is the task of the data management system of the device to resolve such conflicts without interference from a user. In this paper, we take a first step in the development of a probabilistic XML DBMS. The main idea is to drop the assumption that data in the database should be certain: subtrees in XML documents may denote possible views on the real world. We formally define the notion of probabilistic XML tree and several operations thereon. We also present an approach for determining a logical semantics for queries on probabilistic XML data. Finally, we introduce an approach for XML data integration where conflicts are resolved by the introduction of possibilities in the database.","PeriodicalId":297231,"journal":{"name":"21st International Conference on Data Engineering (ICDE'05)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123605581","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 125
Efficient creation and incremental maintenance of the HOPI index for complex XML document collections 为复杂的 XML 文档集高效创建和增量维护 HOPI 索引
Pub Date : 2005-04-05 DOI: 10.1109/ICDE.2005.57
Ralf Schenkel, A. Theobald, G. Weikum
The HOPI index, a connection index for XML documents based on the concept of a 2-hop cover, provides space- and time-efficient reachability tests along the ancestor, descendant, and link axes to support path expressions with wildcards in XML search engines. This paper presents enhanced algorithms for building HOPI, shows how to augment the index with distance information, and discusses incremental index maintenance. Our experiments show substantial improvements over the existing divide-and-conquer algorithm for index creation, low space overhead for including distance information in the index, and efficient updates.
HOPI索引是基于2跳覆盖概念的XML文档的连接索引,它沿着祖先轴、后代轴和链接轴提供空间和时间效率高的可达性测试,以支持XML搜索引擎中带有通配符的路径表达式。本文提出了构建HOPI的增强算法,展示了如何使用距离信息增加索引,并讨论了增量索引维护。我们的实验表明,在索引创建方面,该算法比现有的分治算法有了很大的改进,在索引中包含距离信息的空间开销低,更新效率高。
{"title":"Efficient creation and incremental maintenance of the HOPI index for complex XML document collections","authors":"Ralf Schenkel, A. Theobald, G. Weikum","doi":"10.1109/ICDE.2005.57","DOIUrl":"https://doi.org/10.1109/ICDE.2005.57","url":null,"abstract":"The HOPI index, a connection index for XML documents based on the concept of a 2-hop cover, provides space- and time-efficient reachability tests along the ancestor, descendant, and link axes to support path expressions with wildcards in XML search engines. This paper presents enhanced algorithms for building HOPI, shows how to augment the index with distance information, and discusses incremental index maintenance. Our experiments show substantial improvements over the existing divide-and-conquer algorithm for index creation, low space overhead for including distance information in the index, and efficient updates.","PeriodicalId":297231,"journal":{"name":"21st International Conference on Data Engineering (ICDE'05)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127940546","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 121
Batched processing for information filters 信息过滤器的批处理
Pub Date : 2005-04-05 DOI: 10.1109/ICDE.2005.25
Peter M. Fischer, Donald Kossmann
This paper describes batching, a novel technique in order to improve the throughput of an information filter (e.g. message broker or publish & subscribe system). Rather than processing each message individually, incoming messages are reordered, grouped and a whole group of similar messages is processed. This paper presents alternative strategies to do batching. Extensive performance experiments are conducted on those strategies in order to compare their tradeoffs.
为了提高信息过滤器(如消息代理或发布&订阅系统)的吞吐量,本文描述了一种新的技术——批处理技术。传入的消息不是单独处理每条消息,而是重新排序、分组,并处理一整组类似的消息。本文提出了批处理的替代策略。对这些策略进行了广泛的性能实验,以比较它们的权衡。
{"title":"Batched processing for information filters","authors":"Peter M. Fischer, Donald Kossmann","doi":"10.1109/ICDE.2005.25","DOIUrl":"https://doi.org/10.1109/ICDE.2005.25","url":null,"abstract":"This paper describes batching, a novel technique in order to improve the throughput of an information filter (e.g. message broker or publish & subscribe system). Rather than processing each message individually, incoming messages are reordered, grouped and a whole group of similar messages is processed. This paper presents alternative strategies to do batching. Extensive performance experiments are conducted on those strategies in order to compare their tradeoffs.","PeriodicalId":297231,"journal":{"name":"21st International Conference on Data Engineering (ICDE'05)","volume":"529 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115369502","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Stabbing the sky: efficient skyline computation over sliding windows 刺穿天空:滑动窗口上高效的天际线计算
Pub Date : 2005-04-05 DOI: 10.1109/ICDE.2005.137
Xuemin Lin, Yidong Yuan, Wei Wang, Hongjun Lu
We consider the problem of efficiently computing the skyline against the most recent N elements in a data stream seen so far. Specifically, we study the n-of-N skyline queries; that is, computing the skyline for the most recent n (/spl forall/n/spl les/N) elements. Firstly, we developed an effective pruning technique to minimize the number of elements to be kept. It can be shown that on average storing only O(log/sup d/ N) elements from the most recent N elements is sufficient to support the precise computation of all n-of-N skyline queries in a d-dimension space if the data distribution on each dimension is independent. Then, a novel encoding scheme is proposed, together with efficient update techniques, for the stored elements, so that computing an n-of-N skyline query in a d-dimension space takes O(log N+s) time that is reduced to O(d log log N+s) if the data distribution is independent, where s is the number of skyline points. Thirdly, a novel trigger based technique is provided to process continuous n-of-N skyline queries with O(/spl delta/) time to update the current result per new data element and O(log s) time to update the trigger list per result change, where /spl delta/ is the number of element changes from the current result to the new result. Finally, we extend our techniques to computing the skyline against an arbitrary window in the most recent N element. Besides theoretical performance guarantees, our extensive experiments demonstrated that the new techniques can support on-line skyline query computation over very rapid data streams.
我们考虑的问题是有效地计算天际线对最近的N个元素在数据流中看到到目前为止。具体来说,我们研究了n (n)个skyline查询;也就是说,计算最近n个元素的天际线(所有/n/spl元素/n个元素/spl)。首先,我们开发了一种有效的修剪技术,以尽量减少需要保留的元素数量。可以证明,如果每个维度上的数据分布是独立的,则平均仅存储最近N个元素中的O(log/sup d/ N)个元素足以支持d维空间中所有N (N)个skyline查询的精确计算。然后,对存储的元素提出了一种新的编码方案和高效的更新技术,使得在d维空间中计算N (N)次天际线查询需要O(log N+s)时间,如果数据分布是独立的,则需要O(d log log N+s)时间,其中s为天际线点的个数。第三,提出了一种新颖的基于触发器的技术来处理连续的n-of-N天际线查询,每次新数据元素更新当前结果的时间为O(/spl delta/),每次结果变化更新触发列表的时间为O(log s),其中/spl delta/为从当前结果到新结果的元素变化的次数。最后,我们将我们的技术扩展到针对最近N元素的任意窗口计算天际线。除了理论上的性能保证外,我们的大量实验表明,新技术可以在非常快速的数据流上支持在线天际线查询计算。
{"title":"Stabbing the sky: efficient skyline computation over sliding windows","authors":"Xuemin Lin, Yidong Yuan, Wei Wang, Hongjun Lu","doi":"10.1109/ICDE.2005.137","DOIUrl":"https://doi.org/10.1109/ICDE.2005.137","url":null,"abstract":"We consider the problem of efficiently computing the skyline against the most recent N elements in a data stream seen so far. Specifically, we study the n-of-N skyline queries; that is, computing the skyline for the most recent n (/spl forall/n/spl les/N) elements. Firstly, we developed an effective pruning technique to minimize the number of elements to be kept. It can be shown that on average storing only O(log/sup d/ N) elements from the most recent N elements is sufficient to support the precise computation of all n-of-N skyline queries in a d-dimension space if the data distribution on each dimension is independent. Then, a novel encoding scheme is proposed, together with efficient update techniques, for the stored elements, so that computing an n-of-N skyline query in a d-dimension space takes O(log N+s) time that is reduced to O(d log log N+s) if the data distribution is independent, where s is the number of skyline points. Thirdly, a novel trigger based technique is provided to process continuous n-of-N skyline queries with O(/spl delta/) time to update the current result per new data element and O(log s) time to update the trigger list per result change, where /spl delta/ is the number of element changes from the current result to the new result. Finally, we extend our techniques to computing the skyline against an arbitrary window in the most recent N element. Besides theoretical performance guarantees, our extensive experiments demonstrated that the new techniques can support on-line skyline query computation over very rapid data streams.","PeriodicalId":297231,"journal":{"name":"21st International Conference on Data Engineering (ICDE'05)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116693829","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 290
DCbot: exploring the Web as value-added service for location-based applications DCbot:将Web作为基于位置的应用程序的增值服务进行探索
Pub Date : 2005-04-05 DOI: 10.1109/ICDE.2005.46
M. Jakob, M. Großmann, Nicola Hönle, D. Nicklas
Location-based services (LBS) are typically mobile applications that adapt their behavior to the spatial context of the user, e.g. by providing maps and navigational information of the user's current position. Existing location-based applications rely on spatial data that is gathered and preprocessed especially for them and that is stored by particular data providers. Location-based applications can benefit from World Wide Web and additional information source, if, in a preprocessing step, Web pages are mapped to locations. A model for this is virtual information towers (VIT), spatial Web portals with a location and a visibility area that represents the region where the information is relevant. DCbot processes HTML pages in the WWW like a crawler of a search engine. It analyses the pages using pre-defined rules and spatial knowledge and maps them to locations.
基于位置的服务(LBS)是典型的移动应用程序,可以根据用户的空间环境调整其行为,例如,通过提供用户当前位置的地图和导航信息。现有的基于位置的应用程序依赖于空间数据,这些数据是专门为它们收集和预处理的,并且由特定的数据提供者存储。如果在预处理步骤中将Web页面映射到位置,基于位置的应用程序可以从万维网和其他信息源中获益。这方面的一个模型是虚拟信息塔(VIT),它是具有位置和可见性区域的空间Web门户,表示与信息相关的区域。DCbot像搜索引擎的爬虫一样处理WWW中的HTML页面。它使用预定义的规则和空间知识分析页面,并将它们映射到位置。
{"title":"DCbot: exploring the Web as value-added service for location-based applications","authors":"M. Jakob, M. Großmann, Nicola Hönle, D. Nicklas","doi":"10.1109/ICDE.2005.46","DOIUrl":"https://doi.org/10.1109/ICDE.2005.46","url":null,"abstract":"Location-based services (LBS) are typically mobile applications that adapt their behavior to the spatial context of the user, e.g. by providing maps and navigational information of the user's current position. Existing location-based applications rely on spatial data that is gathered and preprocessed especially for them and that is stored by particular data providers. Location-based applications can benefit from World Wide Web and additional information source, if, in a preprocessing step, Web pages are mapped to locations. A model for this is virtual information towers (VIT), spatial Web portals with a location and a visibility area that represents the region where the information is relevant. DCbot processes HTML pages in the WWW like a crawler of a search engine. It analyses the pages using pre-defined rules and spatial knowledge and maps them to locations.","PeriodicalId":297231,"journal":{"name":"21st International Conference on Data Engineering (ICDE'05)","volume":"78 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116406445","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Progressive distributed top-k retrieval in peer-to-peer networks 点对点网络中的渐进式分布式top-k检索
Pub Date : 2005-04-05 DOI: 10.1109/ICDE.2005.115
Wolf-Tilo Balke, W. Nejdl, W. Siberski, U. Thaden
Query processing in traditional information management systems has moved from an exact match model to more flexible paradigms allowing cooperative retrieval by aggregating the database objects' degree of match for each different query predicate and returning the best matching objects only. In peer-to-peer systems such strategies are even more important, given the potentially large number of peers, which may contribute to the results. Yet current peer-to-peer research has barely started to investigate such approaches. In this paper we discuss the benefits of best match/top-k queries in the context of distributed peer-to-peer information infrastructures and show how to extend the limited query processing in current peer-to-peer networks by allowing the distributed processing of top-k queries, while maintaining a minimum of data traffic. Relying on a super-peer backbone organized in the HyperCuP topology we show how to use local indexes for optimizing the necessary query routing and how to process intermediate results in inner network nodes at the earliest possible point in time cutting down the necessary data traffic within the network. Our algorithm is based on dynamically collected query statistics only, no continuous index update processes are necessary, allowing it to scale easily to large numbers of peers, as well as dynamic additions/deletions of peers. We show our approach to always deliver correct result sets and to be optimal in terms of necessary object accesses and data traffic. Finally, we present simulation results for both static and dynamic network environments.
传统信息管理系统中的查询处理已经从精确匹配模型转向更灵活的范式,通过聚合数据库对象对每个不同查询谓词的匹配程度,并只返回最匹配的对象,从而允许协作检索。在点对点系统中,这种策略甚至更为重要,因为潜在的大量对等体可能对结果有所贡献。然而,目前的点对点研究几乎没有开始调查这些方法。在本文中,我们讨论了最佳匹配/top-k查询在分布式点对点信息基础设施中的好处,并展示了如何通过允许top-k查询的分布式处理来扩展当前点对点网络中有限的查询处理,同时保持最小的数据流量。通过HyperCuP拓扑中组织的超级对等主干,我们展示了如何使用本地索引来优化必要的查询路由,以及如何在尽可能早的时间点处理内部网络节点中的中间结果,从而减少网络中必要的数据流量。我们的算法仅基于动态收集的查询统计信息,不需要连续的索引更新过程,这使得它可以轻松扩展到大量的对等节点,以及动态添加/删除对等节点。我们展示了我们的方法始终提供正确的结果集,并在必要的对象访问和数据流量方面达到最佳。最后,给出了静态和动态网络环境下的仿真结果。
{"title":"Progressive distributed top-k retrieval in peer-to-peer networks","authors":"Wolf-Tilo Balke, W. Nejdl, W. Siberski, U. Thaden","doi":"10.1109/ICDE.2005.115","DOIUrl":"https://doi.org/10.1109/ICDE.2005.115","url":null,"abstract":"Query processing in traditional information management systems has moved from an exact match model to more flexible paradigms allowing cooperative retrieval by aggregating the database objects' degree of match for each different query predicate and returning the best matching objects only. In peer-to-peer systems such strategies are even more important, given the potentially large number of peers, which may contribute to the results. Yet current peer-to-peer research has barely started to investigate such approaches. In this paper we discuss the benefits of best match/top-k queries in the context of distributed peer-to-peer information infrastructures and show how to extend the limited query processing in current peer-to-peer networks by allowing the distributed processing of top-k queries, while maintaining a minimum of data traffic. Relying on a super-peer backbone organized in the HyperCuP topology we show how to use local indexes for optimizing the necessary query routing and how to process intermediate results in inner network nodes at the earliest possible point in time cutting down the necessary data traffic within the network. Our algorithm is based on dynamically collected query statistics only, no continuous index update processes are necessary, allowing it to scale easily to large numbers of peers, as well as dynamic additions/deletions of peers. We show our approach to always deliver correct result sets and to be optimal in terms of necessary object accesses and data traffic. Finally, we present simulation results for both static and dynamic network environments.","PeriodicalId":297231,"journal":{"name":"21st International Conference on Data Engineering (ICDE'05)","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128617061","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 209
Improving preemptive prioritization via statistical characterization of OLTP locking 通过OLTP锁定的统计特性改进抢占优先级
Pub Date : 2005-04-05 DOI: 10.1109/ICDE.2005.78
David T. McWherter, Bianca Schroeder, A. Ailamaki, Mor Harchol-Balter
OLTP and transactional workloads are increasingly common in computer systems, ranging from e-commerce to warehousing to inventory management. It is valuable to provide priority scheduling in these systems, to reduce the response time for the most important clients, e.g. the "big spenders". Two-phase locking, commonly used in DBMS, makes prioritization difficult, as transactions wait for locks held by others regardless of priority. Common lock scheduling solutions, including non-preemptive priority inheritance and preemptive abort, have performance drawbacks for TPC-C type workloads. The contributions of this paper are two-fold: (i) We provide a detailed statistical analysis of locking in TPC-C workloads with priorities under several common preemptive and non-preemptive lock prioritization policies. We determine why non-preemptive policies fail to sufficiently help high-priority transactions, and why preemptive policies excessively hurt low-priority transactions, (ii) We propose and implement a policy, POW, that provides all the benefits of preemptive prioritization without its penalties.
OLTP和事务性工作负载在计算机系统中越来越普遍,从电子商务到仓储再到库存管理。在这些系统中提供优先级调度是有价值的,可以减少最重要客户端的响应时间,例如“大客户”。通常在DBMS中使用的两阶段锁定使得优先级难以确定,因为事务等待其他事务持有的锁,而不管优先级如何。常见的锁调度解决方案,包括非抢占式优先级继承和抢占式中止,对于TPC-C类型的工作负载都存在性能缺陷。本文的贡献有两个方面:(i)我们在几种常见的抢占式和非抢占式锁优先级策略下,对具有优先级的TPC-C工作负载中的锁定进行了详细的统计分析。我们确定了为什么非抢占策略不能充分帮助高优先级事务,以及为什么抢占策略过度伤害低优先级事务,(ii)我们提出并实施了一个策略,POW,它提供了抢占优先级的所有好处而没有惩罚。
{"title":"Improving preemptive prioritization via statistical characterization of OLTP locking","authors":"David T. McWherter, Bianca Schroeder, A. Ailamaki, Mor Harchol-Balter","doi":"10.1109/ICDE.2005.78","DOIUrl":"https://doi.org/10.1109/ICDE.2005.78","url":null,"abstract":"OLTP and transactional workloads are increasingly common in computer systems, ranging from e-commerce to warehousing to inventory management. It is valuable to provide priority scheduling in these systems, to reduce the response time for the most important clients, e.g. the \"big spenders\". Two-phase locking, commonly used in DBMS, makes prioritization difficult, as transactions wait for locks held by others regardless of priority. Common lock scheduling solutions, including non-preemptive priority inheritance and preemptive abort, have performance drawbacks for TPC-C type workloads. The contributions of this paper are two-fold: (i) We provide a detailed statistical analysis of locking in TPC-C workloads with priorities under several common preemptive and non-preemptive lock prioritization policies. We determine why non-preemptive policies fail to sufficiently help high-priority transactions, and why preemptive policies excessively hurt low-priority transactions, (ii) We propose and implement a policy, POW, that provides all the benefits of preemptive prioritization without its penalties.","PeriodicalId":297231,"journal":{"name":"21st International Conference on Data Engineering (ICDE'05)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121914001","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 32
Adlib: a self-tuning index for dynamic peer-to-peer systems Adlib:动态对等系统的自调优索引
Pub Date : 2005-04-05 DOI: 10.1109/ICDE.2005.19
Prasanna Ganesan, Qixiang Sun, H. Garcia-Molina
Peer-to-peer (P2P) systems enable queries over a large database horizontally partitioned across a dynamic set of nodes. We devise a self-tuning index for such systems that can trade off index maintenance cost against query efficiency, in order to optimize the overall system cost. The index, Adlib, dynamically adapts itself to operate at the optimal trade-off point, even as the optimal configuration changes with nodes joining and leaving the system. We use experiments on realistic workloads to demonstrate that Adlib can reduce the overall system cost by a factor of four.
点对点(P2P)系统支持对跨一组动态节点水平分区的大型数据库进行查询。我们为这样的系统设计了一个自调优索引,它可以权衡索引维护成本和查询效率,以优化总体系统成本。索引Adlib动态地调整自身以在最佳折衷点运行,即使最优配置随着节点加入和离开系统而改变。我们使用实际工作负载上的实验来证明Adlib可以将整个系统成本降低四倍。
{"title":"Adlib: a self-tuning index for dynamic peer-to-peer systems","authors":"Prasanna Ganesan, Qixiang Sun, H. Garcia-Molina","doi":"10.1109/ICDE.2005.19","DOIUrl":"https://doi.org/10.1109/ICDE.2005.19","url":null,"abstract":"Peer-to-peer (P2P) systems enable queries over a large database horizontally partitioned across a dynamic set of nodes. We devise a self-tuning index for such systems that can trade off index maintenance cost against query efficiency, in order to optimize the overall system cost. The index, Adlib, dynamically adapts itself to operate at the optimal trade-off point, even as the optimal configuration changes with nodes joining and leaving the system. We use experiments on realistic workloads to demonstrate that Adlib can reduce the overall system cost by a factor of four.","PeriodicalId":297231,"journal":{"name":"21st International Conference on Data Engineering (ICDE'05)","volume":"103 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117315924","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
Triggers over XML views of relational data 关系数据的XML视图上的触发器
Pub Date : 2005-04-05 DOI: 10.1109/ICDE.2005.147
F. Shao, Antal F. Novak, J. Shanmugasundaram
XML has emerged as a dominant standard for information exchange on the Internet. However, a large fraction of data continues to be stored in relational databases. At a high level, there are two approaches to supporting triggers over XML views. The first is to materialize the entire view and store it in an XML database with support for XML triggers. However, this approach suffers from the overhead of replicating and incrementally maintaining the materialized XML on every relational update affecting the view, even though users may only be interested in relatively rare events. In this paper, we propose the alternative approach of translating XML triggers into SQL triggers. There are some challenges involved in this approach, however, because triggers can be specified over complex XML views with nested predicates, while SQL triggers can only be specified over flat tables. Consequently, even identifying the parts of an XML view that could have changed due to a (possibly deeply nested) SQL update is a non-trivial task, as is the problem of computing the old and new values of an updated fragment of the view. We address the above challenges and propose a system architecture and an algorithm for supporting triggers over XML views of relational data. We implement and evaluate our system; the performance results indicate our techniques are a feasible approach to supporting triggers over XML views of relational data.
XML已经成为Internet上信息交换的主要标准。然而,很大一部分数据仍然存储在关系数据库中。在高层次上,有两种方法支持XML视图上的触发器。第一种方法是实现整个视图并将其存储在支持XML触发器的XML数据库中。但是,即使用户可能只对相对罕见的事件感兴趣,这种方法也有在影响视图的每个关系更新上复制和增量维护物化XML的开销。在本文中,我们提出了将XML触发器转换为SQL触发器的替代方法。但是,这种方法存在一些挑战,因为可以在带有嵌套谓词的复杂XML视图上指定触发器,而SQL触发器只能在平面表上指定。因此,即使识别由于SQL更新(可能是深度嵌套的)而可能发生变化的XML视图部分也是一项重要的任务,计算视图更新片段的新旧值的问题也是如此。我们解决了上述挑战,并提出了一种系统架构和算法,用于支持关系数据的XML视图上的触发器。我们实施和评估我们的系统;性能结果表明,我们的技术是在关系数据的XML视图上支持触发器的可行方法。
{"title":"Triggers over XML views of relational data","authors":"F. Shao, Antal F. Novak, J. Shanmugasundaram","doi":"10.1109/ICDE.2005.147","DOIUrl":"https://doi.org/10.1109/ICDE.2005.147","url":null,"abstract":"XML has emerged as a dominant standard for information exchange on the Internet. However, a large fraction of data continues to be stored in relational databases. At a high level, there are two approaches to supporting triggers over XML views. The first is to materialize the entire view and store it in an XML database with support for XML triggers. However, this approach suffers from the overhead of replicating and incrementally maintaining the materialized XML on every relational update affecting the view, even though users may only be interested in relatively rare events. In this paper, we propose the alternative approach of translating XML triggers into SQL triggers. There are some challenges involved in this approach, however, because triggers can be specified over complex XML views with nested predicates, while SQL triggers can only be specified over flat tables. Consequently, even identifying the parts of an XML view that could have changed due to a (possibly deeply nested) SQL update is a non-trivial task, as is the problem of computing the old and new values of an updated fragment of the view. We address the above challenges and propose a system architecture and an algorithm for supporting triggers over XML views of relational data. We implement and evaluate our system; the performance results indicate our techniques are a feasible approach to supporting triggers over XML views of relational data.","PeriodicalId":297231,"journal":{"name":"21st International Conference on Data Engineering (ICDE'05)","volume":"209 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123016523","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Load and network aware query routing for information integration 负载和网络感知查询路由信息集成
Pub Date : 2005-04-05 DOI: 10.1109/ICDE.2005.83
Wen-Syan Li, Vishal S. Batra, Vijayshankar Raman, Wei Han, K. Candan, I. Narang
Current federated systems deploy cost-based query optimization mechanisms; i.e., the optimizer selects a global query plan with the lowest cost to execute. Thus, cost functions influence what remote sources (i.e. equivalent data sources) to access and how federated queries are processed. In most federated systems, the underlying cost model is based on database statistics and query statements; however, the system load of remote sources and the dynamic nature of the network latency in wide area networks are not considered. As a result, federated query processing solutions can not adapt to runtime environment changes, such as network congestion or heavy workloads at remote sources. We present a novel system architecture that deploys a query cost calibrator to calibrate the cost function based on system load and network latency at the remote sources and consequently indirectly "influences" query routing and load distribution in federated information systems.
当前的联邦系统部署基于成本的查询优化机制;也就是说,优化器选择执行成本最低的全局查询计划。因此,成本函数影响访问哪些远程数据源(即等效数据源)以及如何处理联邦查询。在大多数联邦系统中,底层成本模型基于数据库统计信息和查询语句;然而,在广域网中,没有考虑远程源的系统负载和网络延迟的动态性。因此,联邦查询处理解决方案不能适应运行时环境的变化,例如网络拥塞或远程数据源上的繁重工作负载。我们提出了一种新的系统架构,它部署了一个查询成本校准器,根据远程源的系统负载和网络延迟来校准成本函数,从而间接“影响”联邦信息系统中的查询路由和负载分配。
{"title":"Load and network aware query routing for information integration","authors":"Wen-Syan Li, Vishal S. Batra, Vijayshankar Raman, Wei Han, K. Candan, I. Narang","doi":"10.1109/ICDE.2005.83","DOIUrl":"https://doi.org/10.1109/ICDE.2005.83","url":null,"abstract":"Current federated systems deploy cost-based query optimization mechanisms; i.e., the optimizer selects a global query plan with the lowest cost to execute. Thus, cost functions influence what remote sources (i.e. equivalent data sources) to access and how federated queries are processed. In most federated systems, the underlying cost model is based on database statistics and query statements; however, the system load of remote sources and the dynamic nature of the network latency in wide area networks are not considered. As a result, federated query processing solutions can not adapt to runtime environment changes, such as network congestion or heavy workloads at remote sources. We present a novel system architecture that deploys a query cost calibrator to calibrate the cost function based on system load and network latency at the remote sources and consequently indirectly \"influences\" query routing and load distribution in federated information systems.","PeriodicalId":297231,"journal":{"name":"21st International Conference on Data Engineering (ICDE'05)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115049141","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
期刊
21st International Conference on Data Engineering (ICDE'05)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1