首页 > 最新文献

2016 IEEE 32nd International Conference on Data Engineering (ICDE)最新文献

英文 中文
Leveraging non-volatile memory for instant restarts of in-memory database systems 利用非易失性内存立即重新启动内存中的数据库系统
Pub Date : 2016-05-16 DOI: 10.1109/ICDE.2016.7498351
David Schwalb, Martin Faust, Markus Dreseler, Pedro Flemming, H. Plattner
Emerging non-volatile memory technologies (NVM) offer fast and byte-addressable access, allowing to rethink the durability mechanisms of in-memory databases. Hyrise-NV is a database storage engine that maintains table and index structures on NVM. Our architecture updates the database state and index structures transactionally consistent on NVM using multi-version data structures, allowing to instantly recover data-bases independent of their size. In this paper, we demonstrate the instant restart capabilities of Hyrise-NV, storing all data on non-volatile memory. Recovering a dataset of size 92.2 GB takes about 53 seconds using our log-based approach, whereas Hyrise-NV recovers in under one second.
新兴的非易失性内存技术(NVM)提供了快速且可字节寻址的访问,允许重新考虑内存数据库的持久性机制。Hyrise-NV是一个数据库存储引擎,用于在NVM上维护表和索引结构。我们的架构使用多版本数据结构在NVM上以事务一致的方式更新数据库状态和索引结构,允许立即恢复与数据库大小无关的数据库。在本文中,我们演示了Hyrise-NV的即时重启能力,将所有数据存储在非易失性存储器上。使用基于日志的方法恢复大小为92.2 GB的数据集大约需要53秒,而Hyrise-NV的恢复时间不到1秒。
{"title":"Leveraging non-volatile memory for instant restarts of in-memory database systems","authors":"David Schwalb, Martin Faust, Markus Dreseler, Pedro Flemming, H. Plattner","doi":"10.1109/ICDE.2016.7498351","DOIUrl":"https://doi.org/10.1109/ICDE.2016.7498351","url":null,"abstract":"Emerging non-volatile memory technologies (NVM) offer fast and byte-addressable access, allowing to rethink the durability mechanisms of in-memory databases. Hyrise-NV is a database storage engine that maintains table and index structures on NVM. Our architecture updates the database state and index structures transactionally consistent on NVM using multi-version data structures, allowing to instantly recover data-bases independent of their size. In this paper, we demonstrate the instant restart capabilities of Hyrise-NV, storing all data on non-volatile memory. Recovering a dataset of size 92.2 GB takes about 53 seconds using our log-based approach, whereas Hyrise-NV recovers in under one second.","PeriodicalId":6883,"journal":{"name":"2016 IEEE 32nd International Conference on Data Engineering (ICDE)","volume":"1 1","pages":"1386-1389"},"PeriodicalIF":0.0,"publicationDate":"2016-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88599417","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Quality-driven disorder handling for m-way sliding window stream joins m-way滑动窗口流连接的质量驱动无序处理
Pub Date : 2016-05-16 DOI: 10.1109/ICDE.2016.7498265
Yuanzhen Ji, Jun Sun, A. Nica, Zbigniew Jerzak, Gregor Hackenbroich, C. Fetzer
Sliding window join is one of the most important operators for stream applications. To produce high quality join results, a stream processing system must deal with the ubiquitous disorder within input streams which is caused by network delay, parallel processing, etc. Disorder handling involves an inevitable tradeoff between the latency and the quality of produced join results. To meet different requirements of stream applications, it is desirable to provide a user-configurable result-latency vs. result-quality tradeoff. Existing disorder handling approaches either do not provide such configurability, or support only user-specified latency constraints. In this work, we advocate the idea of quality-driven disorder handling, and propose a buffer-based disorder handling approach for sliding window joins, which minimizes sizes of input-sorting buffers, thus the result latency, while respecting user-specified result-quality requirements. The core of our approach is an analytical model which directly captures the relationship between sizes of input buffers and the produced result quality. Our approach is generic. It supports m-way sliding window joins with arbitrary join conditions. Experiments on real-world and synthetic datasets show that, compared to the state of the art, our approach can reduce the result latency incurred by disorder handling by up to 95% while providing the same level of result quality.
滑动窗口连接是流应用程序中最重要的操作符之一。为了产生高质量的连接结果,流处理系统必须处理由网络延迟、并行处理等引起的输入流中普遍存在的无序。无序处理涉及在延迟和生成的连接结果质量之间进行不可避免的权衡。为了满足流应用程序的不同需求,最好提供用户可配置的结果延迟与结果质量之间的权衡。现有的混乱处理方法要么不提供这种可配置性,要么只支持用户指定的延迟约束。在这项工作中,我们提倡质量驱动的无序处理思想,并提出了一种基于缓冲区的滑动窗口连接的无序处理方法,该方法最小化了输入排序缓冲区的大小,从而减少了结果延迟,同时尊重用户指定的结果质量要求。我们方法的核心是一个分析模型,它直接捕捉输入缓冲区大小和生成结果质量之间的关系。我们的方法是通用的。它支持任意连接条件的m-way滑动窗口连接。在真实世界和合成数据集上的实验表明,与目前的技术水平相比,我们的方法可以在提供相同水平的结果质量的同时,将无序处理引起的结果延迟减少高达95%。
{"title":"Quality-driven disorder handling for m-way sliding window stream joins","authors":"Yuanzhen Ji, Jun Sun, A. Nica, Zbigniew Jerzak, Gregor Hackenbroich, C. Fetzer","doi":"10.1109/ICDE.2016.7498265","DOIUrl":"https://doi.org/10.1109/ICDE.2016.7498265","url":null,"abstract":"Sliding window join is one of the most important operators for stream applications. To produce high quality join results, a stream processing system must deal with the ubiquitous disorder within input streams which is caused by network delay, parallel processing, etc. Disorder handling involves an inevitable tradeoff between the latency and the quality of produced join results. To meet different requirements of stream applications, it is desirable to provide a user-configurable result-latency vs. result-quality tradeoff. Existing disorder handling approaches either do not provide such configurability, or support only user-specified latency constraints. In this work, we advocate the idea of quality-driven disorder handling, and propose a buffer-based disorder handling approach for sliding window joins, which minimizes sizes of input-sorting buffers, thus the result latency, while respecting user-specified result-quality requirements. The core of our approach is an analytical model which directly captures the relationship between sizes of input buffers and the produced result quality. Our approach is generic. It supports m-way sliding window joins with arbitrary join conditions. Experiments on real-world and synthetic datasets show that, compared to the state of the art, our approach can reduce the result latency incurred by disorder handling by up to 95% while providing the same level of result quality.","PeriodicalId":6883,"journal":{"name":"2016 IEEE 32nd International Conference on Data Engineering (ICDE)","volume":"7 1","pages":"493-504"},"PeriodicalIF":0.0,"publicationDate":"2016-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85935076","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Edge classification in networks 网络中的边缘分类
Pub Date : 2016-05-16 DOI: 10.1109/ICDE.2016.7498311
C. Aggarwal, Gewen He, Peixiang Zhao
We consider in this paper the edge classification problem in networks, which is defined as follows. Given a graph-structured network G(N, A), where N is a set of vertices and A ⊆ N ×N is a set of edges, in which a subset Al ⊆ A of edges are properly labeled a priori, determine for those edges in Au = AAl the edge labels which are unknown. The edge classification problem has numerous applications in graph mining and social network analysis, such as relationship discovery, categorization, and recommendation. Although the vertex classification problem has been well known and extensively explored in networks, edge classification is relatively unknown and in an urgent need for careful studies. In this paper, we present a series of efficient, neighborhood-based algorithms to perform edge classification in networks. To make the proposed algorithms scalable in large-scale networks, which can be either disk-resident or streamlike, we further devise efficient, cost-effective probabilistic edge classification methods without a significant compromise to the classification accuracy. We carry out experimental studies in a series of real-world networks, and the experimental results demonstrate both the effectiveness and efficiency of the proposed methods for edge classification in large networks.
本文考虑网络中的边缘分类问题,定义如下:给定一个图结构网络G(N, a),其中N为一组顶点,a≥×N为一组边,其中一个子集Al≥a的边被先验地适当标记,在Au = a Al中确定未知边的标记。边缘分类问题在图挖掘和社会网络分析中有许多应用,如关系发现、分类和推荐。虽然顶点分类问题已经在网络中得到了广泛的研究,但边缘分类问题相对来说还是一个未知的问题,亟待深入研究。在本文中,我们提出了一系列有效的,基于邻域的算法来执行网络中的边缘分类。为了使所提出的算法在磁盘驻留或流状的大规模网络中具有可扩展性,我们进一步设计了高效,成本效益高的概率边缘分类方法,而不会显著损害分类精度。我们在一系列现实网络中进行了实验研究,实验结果证明了所提出的方法在大型网络中边缘分类的有效性和效率。
{"title":"Edge classification in networks","authors":"C. Aggarwal, Gewen He, Peixiang Zhao","doi":"10.1109/ICDE.2016.7498311","DOIUrl":"https://doi.org/10.1109/ICDE.2016.7498311","url":null,"abstract":"We consider in this paper the edge classification problem in networks, which is defined as follows. Given a graph-structured network G(N, A), where N is a set of vertices and A ⊆ N ×N is a set of edges, in which a subset Al ⊆ A of edges are properly labeled a priori, determine for those edges in Au = AAl the edge labels which are unknown. The edge classification problem has numerous applications in graph mining and social network analysis, such as relationship discovery, categorization, and recommendation. Although the vertex classification problem has been well known and extensively explored in networks, edge classification is relatively unknown and in an urgent need for careful studies. In this paper, we present a series of efficient, neighborhood-based algorithms to perform edge classification in networks. To make the proposed algorithms scalable in large-scale networks, which can be either disk-resident or streamlike, we further devise efficient, cost-effective probabilistic edge classification methods without a significant compromise to the classification accuracy. We carry out experimental studies in a series of real-world networks, and the experimental results demonstrate both the effectiveness and efficiency of the proposed methods for edge classification in large networks.","PeriodicalId":6883,"journal":{"name":"2016 IEEE 32nd International Conference on Data Engineering (ICDE)","volume":"5 1","pages":"1038-1049"},"PeriodicalIF":0.0,"publicationDate":"2016-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85067625","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Computing Connected Components with linear communication cost in pregel-like systems 类预凝胶系统中具有线性通信代价的连通组件计算
Pub Date : 2016-05-16 DOI: 10.1109/ICDE.2016.7498231
Xing Feng, Lijun Chang, Xuemin Lin, Lu Qin, W. Zhang
The paper studies two fundamental problems in graph analytics: computing Connected Components (CCs) and computing BiConnected Components (BCCs) of a graph. With the recent advent of Big Data, developing effcient distributed algorithms for computing CCs and BCCs of a big graph has received increasing interests. As with the existing research efforts, in this paper we focus on the Pregel programming model, while the techniques may be extended to other programming models including MapReduce and Spark. The state-of-the-art techniques for computing CCs and BCCs in Pregel incur O(m × #supersteps) total costs for both data communication and computation, where m is the number of edges in a graph and #supersteps is the number of supersteps. Since the network communication speed is usually much slower than the computation speed, communication costs are the dominant costs of the total running time in the existing techniques. In this paper, we propose a new paradigm based on graph decomposition to reduce the total communication costs from O(m×#supersteps) to O(m), for both computing CCs and computing BCCs. Moreover, the total computation costs of our techniques are smaller than that of the existing techniques in practice, though theoretically they are almost the same. Comprehensive empirical studies demonstrate that our approaches can outperform the existing techniques by one order of magnitude regarding the total running time.
本文研究了图分析中的两个基本问题:图的连通分量(cc)和双连通分量(bcc)。随着近年来大数据的出现,开发高效的分布式算法来计算大图的cc和bcc受到越来越多的关注。与现有的研究成果一样,本文主要关注Pregel编程模型,而这些技术可以扩展到其他编程模型,包括MapReduce和Spark。在Pregel中,计算cc和bcc的最先进技术在数据通信和计算方面需要O(m × #supersteps)的总成本,其中m是图中边的数量,#supersteps是超步骤的数量。由于网络通信速度通常比计算速度慢得多,因此在现有技术中,通信成本是总运行时间的主要成本。在本文中,我们提出了一种基于图分解的新范式,将计算cc和计算bcc的总通信成本从O(mx# supersteps)降低到O(m)。此外,我们的技术的总计算成本比实践中的现有技术要小,尽管理论上它们几乎相同。综合实证研究表明,我们的方法在总运行时间方面比现有技术高出一个数量级。
{"title":"Computing Connected Components with linear communication cost in pregel-like systems","authors":"Xing Feng, Lijun Chang, Xuemin Lin, Lu Qin, W. Zhang","doi":"10.1109/ICDE.2016.7498231","DOIUrl":"https://doi.org/10.1109/ICDE.2016.7498231","url":null,"abstract":"The paper studies two fundamental problems in graph analytics: computing Connected Components (CCs) and computing BiConnected Components (BCCs) of a graph. With the recent advent of Big Data, developing effcient distributed algorithms for computing CCs and BCCs of a big graph has received increasing interests. As with the existing research efforts, in this paper we focus on the Pregel programming model, while the techniques may be extended to other programming models including MapReduce and Spark. The state-of-the-art techniques for computing CCs and BCCs in Pregel incur O(m × #supersteps) total costs for both data communication and computation, where m is the number of edges in a graph and #supersteps is the number of supersteps. Since the network communication speed is usually much slower than the computation speed, communication costs are the dominant costs of the total running time in the existing techniques. In this paper, we propose a new paradigm based on graph decomposition to reduce the total communication costs from O(m×#supersteps) to O(m), for both computing CCs and computing BCCs. Moreover, the total computation costs of our techniques are smaller than that of the existing techniques in practice, though theoretically they are almost the same. Comprehensive empirical studies demonstrate that our approaches can outperform the existing techniques by one order of magnitude regarding the total running time.","PeriodicalId":6883,"journal":{"name":"2016 IEEE 32nd International Conference on Data Engineering (ICDE)","volume":"63 1","pages":"85-96"},"PeriodicalIF":0.0,"publicationDate":"2016-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84335508","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Answering why-not questions on metric probabilistic range queries 回答关于度量概率范围查询的why-not问题
Pub Date : 2016-05-16 DOI: 10.1109/ICDE.2016.7498288
Lu Chen, Yunjun Gao, Kai Wang, Christian S. Jensen, Gang Chen
Metric probabilistic range queries (MPRQ) have received substantial attention due to their utility in multimedia and text retrieval, decision making, etc. Existing MPRQ studies generally aim to improve query efficiency and resource usage. In contrast, we define and offer solutions to why-not questions on MPRQ. Given an original metric probabilistic range query and a why-not set W of uncertain objects that are absent from the query result, a why-not question on MPRQ explains why the uncertain objects in W do not appear in the query result, and provides refinements of the original query and/or W with the minimal penalty, so that the uncertain objects in W appear in the result of the refined query. Specifically, we propose a framework that consists of three efficient solutions, one that modifies the original query, one that modifies the why-not set, and one that modifies both the original query and the why-not set. Extensive experiments using both real and synthetic data sets offer insights into the properties of the proposed algorithms, and show that they are effective and efficient.
度量概率范围查询(MPRQ)由于在多媒体和文本检索、决策等方面的应用而受到了广泛的关注。现有的MPRQ研究一般以提高查询效率和资源利用率为目标。相反,我们定义并提供解决MPRQ中“为什么不”问题的方法。给定一个原始度量概率范围查询和查询结果中不确定对象的why-not集合W, MPRQ上的why-not问题解释了为什么W中的不确定对象没有出现在查询结果中,并以最小的惩罚对原始查询和/或W进行改进,使W中的不确定对象出现在改进后的查询结果中。具体来说,我们提出了一个由三个有效解决方案组成的框架,一个修改原始查询,一个修改为什么不设置,另一个修改原始查询和为什么不设置。使用真实和合成数据集的大量实验提供了对所提出算法特性的见解,并表明它们是有效和高效的。
{"title":"Answering why-not questions on metric probabilistic range queries","authors":"Lu Chen, Yunjun Gao, Kai Wang, Christian S. Jensen, Gang Chen","doi":"10.1109/ICDE.2016.7498288","DOIUrl":"https://doi.org/10.1109/ICDE.2016.7498288","url":null,"abstract":"Metric probabilistic range queries (MPRQ) have received substantial attention due to their utility in multimedia and text retrieval, decision making, etc. Existing MPRQ studies generally aim to improve query efficiency and resource usage. In contrast, we define and offer solutions to why-not questions on MPRQ. Given an original metric probabilistic range query and a why-not set W of uncertain objects that are absent from the query result, a why-not question on MPRQ explains why the uncertain objects in W do not appear in the query result, and provides refinements of the original query and/or W with the minimal penalty, so that the uncertain objects in W appear in the result of the refined query. Specifically, we propose a framework that consists of three efficient solutions, one that modifies the original query, one that modifies the why-not set, and one that modifies both the original query and the why-not set. Extensive experiments using both real and synthetic data sets offer insights into the properties of the proposed algorithms, and show that they are effective and efficient.","PeriodicalId":6883,"journal":{"name":"2016 IEEE 32nd International Conference on Data Engineering (ICDE)","volume":"57 1","pages":"767-778"},"PeriodicalIF":0.0,"publicationDate":"2016-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88035175","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Accelerating database workloads by software-hardware-system co-design 通过软件-硬件系统协同设计加速数据库工作负载
Pub Date : 2016-05-16 DOI: 10.1109/ICDE.2016.7498362
R. Bordawekar, Mohammad Sadoghi
The key objective of this tutorial is to provide a broad, yet an in-depth survey of the emerging field of co-designing software, hardware, and systems components for accelerating enterprise data management workloads. The overall goal of this tutorial is two-fold. First, we provide a concise system-level characterization of different types of data management technologies, namely, the relational and NoSQL databases and data stream management systems from the perspective of analytical workloads. Using the characterization, we discuss opportunities for accelerating key data management workloads using software and hardware approaches. Second, we dive deeper into the hardware acceleration opportunities using Graphics Processing Units (GPUs) and Field-Programmable Gate Arrays (FPGAs) for the query execution pipeline. Furthermore, we explore other hardware acceleration mechanisms such as single-instruction multiple-data (SIMD) that enables short-vector data parallelism.
本教程的主要目标是对用于加速企业数据管理工作负载的共同设计软件、硬件和系统组件这一新兴领域进行广泛而深入的调查。本教程的总体目标有两个。首先,我们从分析工作负载的角度对不同类型的数据管理技术,即关系数据库和NoSQL数据库以及数据流管理系统进行了简明的系统级描述。通过描述,我们讨论了使用软件和硬件方法加速关键数据管理工作负载的机会。其次,我们深入研究了使用图形处理单元(gpu)和现场可编程门阵列(fpga)进行查询执行管道的硬件加速机会。此外,我们还探索了其他硬件加速机制,例如支持短向量数据并行的单指令多数据(SIMD)。
{"title":"Accelerating database workloads by software-hardware-system co-design","authors":"R. Bordawekar, Mohammad Sadoghi","doi":"10.1109/ICDE.2016.7498362","DOIUrl":"https://doi.org/10.1109/ICDE.2016.7498362","url":null,"abstract":"The key objective of this tutorial is to provide a broad, yet an in-depth survey of the emerging field of co-designing software, hardware, and systems components for accelerating enterprise data management workloads. The overall goal of this tutorial is two-fold. First, we provide a concise system-level characterization of different types of data management technologies, namely, the relational and NoSQL databases and data stream management systems from the perspective of analytical workloads. Using the characterization, we discuss opportunities for accelerating key data management workloads using software and hardware approaches. Second, we dive deeper into the hardware acceleration opportunities using Graphics Processing Units (GPUs) and Field-Programmable Gate Arrays (FPGAs) for the query execution pipeline. Furthermore, we explore other hardware acceleration mechanisms such as single-instruction multiple-data (SIMD) that enables short-vector data parallelism.","PeriodicalId":6883,"journal":{"name":"2016 IEEE 32nd International Conference on Data Engineering (ICDE)","volume":"17 1","pages":"1428-1431"},"PeriodicalIF":0.0,"publicationDate":"2016-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87190999","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Context-aware advertisement recommendation for high-speed social news feeding 基于上下文感知的广告推荐,支持高速社交新闻推送
Pub Date : 2016-05-16 DOI: 10.1109/ICDE.2016.7498266
Yuchen Li, Dongxiang Zhang, Ziquan Lan, K. Tan
Social media advertising is a multi-billion dollar market and has become the major revenue source for Facebook and Twitter. To deliver ads to potentially interested users, these social network platforms learn a prediction model for each user based on their personal interests. However, as user interests often evolve slowly, the user may end up receiving repetitive ads. In this paper, we propose a context-aware advertising framework that takes into account the relatively static personal interests as well as the dynamic news feed from friends to drive growth in the ad click-through rate. To meet the real-time requirement, we first propose an online retrieval strategy that finds k most relevant ads matching the dynamic context when a read operation is triggered. To avoid frequent retrieval when the context varies little, we propose a safe region method to quickly determine whether the top-k ads of a user are changed. Finally, we propose a hybrid model to combine the merits of both methods by analyzing the dynamism of news feed to determine an appropriate retrieval strategy. Extensive experiments conducted on multiple real social networks and ad datasets verified the efficiency and robustness of our hybrid model.
社交媒体广告是一个价值数十亿美元的市场,已成为Facebook和Twitter的主要收入来源。为了向潜在感兴趣的用户投放广告,这些社交网络平台根据每个用户的个人兴趣学习了一个预测模型。然而,由于用户兴趣往往演变缓慢,用户最终可能会收到重复的广告。在本文中,我们提出了一个上下文感知广告框架,该框架考虑了相对静态的个人兴趣以及来自朋友的动态新闻馈送,以推动广告点击率的增长。为了满足实时需求,我们首先提出了一种在线检索策略,当读取操作被触发时,该策略可以找到与动态上下文匹配的k个最相关的广告。为了避免上下文变化不大时频繁检索,我们提出了一种安全区域方法来快速确定用户的前k个广告是否发生了变化。最后,我们提出了一个混合模型,通过分析新闻源的动态特性,将两种方法的优点结合起来,以确定合适的检索策略。在多个真实社交网络和广告数据集上进行的大量实验验证了我们的混合模型的效率和鲁棒性。
{"title":"Context-aware advertisement recommendation for high-speed social news feeding","authors":"Yuchen Li, Dongxiang Zhang, Ziquan Lan, K. Tan","doi":"10.1109/ICDE.2016.7498266","DOIUrl":"https://doi.org/10.1109/ICDE.2016.7498266","url":null,"abstract":"Social media advertising is a multi-billion dollar market and has become the major revenue source for Facebook and Twitter. To deliver ads to potentially interested users, these social network platforms learn a prediction model for each user based on their personal interests. However, as user interests often evolve slowly, the user may end up receiving repetitive ads. In this paper, we propose a context-aware advertising framework that takes into account the relatively static personal interests as well as the dynamic news feed from friends to drive growth in the ad click-through rate. To meet the real-time requirement, we first propose an online retrieval strategy that finds k most relevant ads matching the dynamic context when a read operation is triggered. To avoid frequent retrieval when the context varies little, we propose a safe region method to quickly determine whether the top-k ads of a user are changed. Finally, we propose a hybrid model to combine the merits of both methods by analyzing the dynamism of news feed to determine an appropriate retrieval strategy. Extensive experiments conducted on multiple real social networks and ad datasets verified the efficiency and robustness of our hybrid model.","PeriodicalId":6883,"journal":{"name":"2016 IEEE 32nd International Conference on Data Engineering (ICDE)","volume":"43 1","pages":"505-516"},"PeriodicalIF":0.0,"publicationDate":"2016-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86511555","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
MuVE: Efficient Multi-Objective View Recommendation for Visual Data Exploration 面向可视化数据探索的高效多目标视图推荐
Pub Date : 2016-05-16 DOI: 10.1109/ICDE.2016.7498285
Humaira Ehsan, M. Sharaf, Panos K. Chrysanthis
To support effective data exploration, there is a well-recognized need for solutions that can automatically recommend interesting visualizations, which reveal useful insights into the analyzed data. However, such visualizations come at the expense of high data processing costs, where a large number of views are generated to evaluate their usefulness. Those costs are further escalated in the presence of numerical dimensional attributes, due to the potentially large number of possible binning aggregations, which lead to a drastic increase in the number of possible visualizations. To address that challenge, in this paper we propose the MuVE scheme for Multi-Objective View Recommendation for Visual Data Exploration. MuVE introduces a hybrid multi-objective utility function, which captures the impact of binning on the utility of visualizations. Consequently, novel algorithms are proposed for the efficient recommendation of data visualizations that are based on numerical dimensions. The main idea underlying MuVE is to incrementally and progressively assess the different benefits provided by a visualization, which allows an early pruning of a large number of unnecessary operations. Our extensive experimental results show the significant gains provided by our proposed scheme.
为了支持有效的数据探索,有一个公认的解决方案,可以自动推荐有趣的可视化,从而揭示对分析数据的有用见解。然而,这样的可视化是以高昂的数据处理成本为代价的,因为要生成大量的视图来评估它们的有用性。这些成本在存在数值维度属性时进一步升级,因为潜在的大量可能的分组聚合导致可能的可视化数量急剧增加。为了解决这一挑战,本文提出了用于视觉数据探索的多目标视图推荐的MuVE方案。MuVE引入了一个混合的多目标效用函数,它捕获了分组对可视化效用的影响。因此,提出了新的算法,以有效地推荐基于数值维度的数据可视化。MuVE的主要思想是逐步地评估可视化所提供的不同好处,这允许对大量不必要的操作进行早期修剪。广泛的实验结果表明,我们提出的方案提供了显著的收益。
{"title":"MuVE: Efficient Multi-Objective View Recommendation for Visual Data Exploration","authors":"Humaira Ehsan, M. Sharaf, Panos K. Chrysanthis","doi":"10.1109/ICDE.2016.7498285","DOIUrl":"https://doi.org/10.1109/ICDE.2016.7498285","url":null,"abstract":"To support effective data exploration, there is a well-recognized need for solutions that can automatically recommend interesting visualizations, which reveal useful insights into the analyzed data. However, such visualizations come at the expense of high data processing costs, where a large number of views are generated to evaluate their usefulness. Those costs are further escalated in the presence of numerical dimensional attributes, due to the potentially large number of possible binning aggregations, which lead to a drastic increase in the number of possible visualizations. To address that challenge, in this paper we propose the MuVE scheme for Multi-Objective View Recommendation for Visual Data Exploration. MuVE introduces a hybrid multi-objective utility function, which captures the impact of binning on the utility of visualizations. Consequently, novel algorithms are proposed for the efficient recommendation of data visualizations that are based on numerical dimensions. The main idea underlying MuVE is to incrementally and progressively assess the different benefits provided by a visualization, which allows an early pruning of a large number of unnecessary operations. Our extensive experimental results show the significant gains provided by our proposed scheme.","PeriodicalId":6883,"journal":{"name":"2016 IEEE 32nd International Conference on Data Engineering (ICDE)","volume":"41 1","pages":"731-742"},"PeriodicalIF":0.0,"publicationDate":"2016-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79845279","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 55
A column store engine for real-time streaming analytics 用于实时流分析的列存储引擎
Pub Date : 2016-05-16 DOI: 10.1109/ICDE.2016.7498332
Alex Skidanov, Anders J. Papito, A. Prout
This paper describes novel aspects of the column store implemented in the MemSQL database engine and describes the design choices made to support real-time streaming workloads. Column stores have traditionally been restricted to data warehouse scenarios where low latency queries are a secondary goal, and where restricting data ingestion to be offline, batched, append-only, or some combination thereof is acceptable. In contrast, the MemSQL column store implementation treats low latency queries and ongoing writes as first class citizens, with a focus on avoiding interference between read, ingest, update, and storage optimization workloads through the use of fragmented snapshot transactions and optimistic storage reordering. This implementation broadens the range of serviceable column store workloads to include those with more stringent demands on query and data latency, such as those backing operational systems used by adtech, financial services, fraud detection and other real-time or data streaming applications.
本文描述了在MemSQL数据库引擎中实现的列存储的新方面,并描述了为支持实时流工作负载所做的设计选择。列存储传统上仅限于数据仓库场景,在这些场景中,低延迟查询是次要目标,并且可以将数据摄取限制为脱机、批处理、仅追加或其某种组合。相比之下,MemSQL列存储实现将低延迟查询和正在进行的写入视为头等大事,重点是通过使用碎片快照事务和乐观存储重排序来避免读取、摄取、更新和存储优化工作负载之间的干扰。这种实现扩大了可服务列存储工作负载的范围,包括那些对查询和数据延迟有更严格要求的工作负载,例如adtech、金融服务、欺诈检测和其他实时或数据流应用程序使用的后台操作系统。
{"title":"A column store engine for real-time streaming analytics","authors":"Alex Skidanov, Anders J. Papito, A. Prout","doi":"10.1109/ICDE.2016.7498332","DOIUrl":"https://doi.org/10.1109/ICDE.2016.7498332","url":null,"abstract":"This paper describes novel aspects of the column store implemented in the MemSQL database engine and describes the design choices made to support real-time streaming workloads. Column stores have traditionally been restricted to data warehouse scenarios where low latency queries are a secondary goal, and where restricting data ingestion to be offline, batched, append-only, or some combination thereof is acceptable. In contrast, the MemSQL column store implementation treats low latency queries and ongoing writes as first class citizens, with a focus on avoiding interference between read, ingest, update, and storage optimization workloads through the use of fragmented snapshot transactions and optimistic storage reordering. This implementation broadens the range of serviceable column store workloads to include those with more stringent demands on query and data latency, such as those backing operational systems used by adtech, financial services, fraud detection and other real-time or data streaming applications.","PeriodicalId":6883,"journal":{"name":"2016 IEEE 32nd International Conference on Data Engineering (ICDE)","volume":"139 1","pages":"1287-1297"},"PeriodicalIF":0.0,"publicationDate":"2016-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79913183","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Hobbes3: Dynamic generation of variable-length signatures for efficient approximate subsequence mappings 动态生成有效的近似子序列映射的变长签名
Pub Date : 2016-05-16 DOI: 10.1109/ICDE.2016.7498238
Jongik Kim, Chen Li, Xiaohui Xie
Recent advances in DNA sequencing have enabled a flood of sequencing-based applications for studying biology and medicine. A key requirement of these applications is to rapidly and accurately map DNA subsequences to a reference genome. This DNA subsequence mapping problem shares core technical challenges with the similarity query processing problem studied in the database research literature. To solve this problem, existing techniques first extract signatures from a query, then retrieve candidate mapping positions from an index using the extracted signatures, and finally verify the candidate positions. The efficiency of these techniques depends critically on signatures selected from queries, while signature selection relies on an indexing scheme of a reference genome. The q-gram inverted indexing, one of the most widely used indexing schemes, can discover candidate positions quickly, but has the limitation that signatures of queries are restricted to fixed-length q-grams. To address the problem, we propose a flexible way to generate variable-length signatures using a fixed-length q-gram index. The proposed technique groups a few q-grams into a variable-length signature, and generates candidate positions for the variable-length signature using the inverted lists of the q-grams. We also propose a novel dynamic programming algorithm to balance between the filtering power of signatures and the overhead of generating candidate positions for the signatures. Through extensive experiments on both simulated and real genomic data, we show that our technique substantially improves the performance of read mapping in terms of both mapping speed and accuracy.
DNA测序的最新进展使基于测序的应用程序在生物学和医学研究中成为可能。这些应用的一个关键要求是快速准确地将DNA子序列映射到参考基因组。该DNA子序列映射问题与数据库研究文献中研究的相似度查询处理问题具有相同的核心技术挑战。为了解决这个问题,现有技术首先从查询中提取签名,然后使用提取的签名从索引中检索候选映射位置,最后验证候选位置。这些技术的效率主要取决于从查询中选择的签名,而签名选择依赖于参考基因组的索引方案。q-gram倒排索引是目前使用最广泛的索引方案之一,它可以快速发现候选位置,但其缺点是查询的签名仅限于固定长度的q-gram。为了解决这个问题,我们提出了一种灵活的方法来使用固定长度的q-gram索引生成变长签名。该技术将几个q-g分组为变长签名,并使用q-g的倒排表生成变长签名的候选位置。我们还提出了一种新的动态规划算法来平衡签名的过滤能力和为签名生成候选位置的开销。通过对模拟和真实基因组数据的大量实验,我们表明我们的技术在映射速度和精度方面都大大提高了读映射的性能。
{"title":"Hobbes3: Dynamic generation of variable-length signatures for efficient approximate subsequence mappings","authors":"Jongik Kim, Chen Li, Xiaohui Xie","doi":"10.1109/ICDE.2016.7498238","DOIUrl":"https://doi.org/10.1109/ICDE.2016.7498238","url":null,"abstract":"Recent advances in DNA sequencing have enabled a flood of sequencing-based applications for studying biology and medicine. A key requirement of these applications is to rapidly and accurately map DNA subsequences to a reference genome. This DNA subsequence mapping problem shares core technical challenges with the similarity query processing problem studied in the database research literature. To solve this problem, existing techniques first extract signatures from a query, then retrieve candidate mapping positions from an index using the extracted signatures, and finally verify the candidate positions. The efficiency of these techniques depends critically on signatures selected from queries, while signature selection relies on an indexing scheme of a reference genome. The q-gram inverted indexing, one of the most widely used indexing schemes, can discover candidate positions quickly, but has the limitation that signatures of queries are restricted to fixed-length q-grams. To address the problem, we propose a flexible way to generate variable-length signatures using a fixed-length q-gram index. The proposed technique groups a few q-grams into a variable-length signature, and generates candidate positions for the variable-length signature using the inverted lists of the q-grams. We also propose a novel dynamic programming algorithm to balance between the filtering power of signatures and the overhead of generating candidate positions for the signatures. Through extensive experiments on both simulated and real genomic data, we show that our technique substantially improves the performance of read mapping in terms of both mapping speed and accuracy.","PeriodicalId":6883,"journal":{"name":"2016 IEEE 32nd International Conference on Data Engineering (ICDE)","volume":"26 1","pages":"169-180"},"PeriodicalIF":0.0,"publicationDate":"2016-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81746037","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
期刊
2016 IEEE 32nd International Conference on Data Engineering (ICDE)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1