首页 > 最新文献

Proceedings of the 10th ACM International Conference on Distributed and Event-based Systems最新文献

英文 中文
Partial pattern fulfillment and its application in event processing: poster 部分模式实现及其在事件处理中的应用:海报
Suad Sejdovic
Within the last few years, event processing has been gaining a lot of attention as it represents a powerful possibility to establish a real-time monitoring and situation detection. The real-time detection of situations allows a timely reaction and helps to reduce the damage made by harmful situations or increase the benefit of opportunities. The so called situation of interest (SOI) is described as a pattern, which represents a combination of events describing their temporal causality. Instead of just detecting SOIs, we search for possibilities to predict them. As patterns are the basis for reactive event processing, we also want to exploit them for proactive event processing. Therefore, we immerse into the world of pattern management to gain a better understanding of patterns, their structure and expressiveness.
在过去几年中,事件处理已经获得了很多关注,因为它代表了建立实时监控和情况检测的强大可能性。实时检测的情况允许及时的反应,并有助于减少有害情况造成的损害或增加机会的好处。所谓的兴趣情境(SOI)被描述为一种模式,它代表了描述其时间因果关系的事件组合。我们不只是探测soi,而是寻找预测它们的可能性。由于模式是响应式事件处理的基础,我们也希望利用它们进行主动事件处理。因此,我们沉浸在模式管理的世界中,以便更好地理解模式、模式的结构和表达性。
{"title":"Partial pattern fulfillment and its application in event processing: poster","authors":"Suad Sejdovic","doi":"10.1145/2933267.2933537","DOIUrl":"https://doi.org/10.1145/2933267.2933537","url":null,"abstract":"Within the last few years, event processing has been gaining a lot of attention as it represents a powerful possibility to establish a real-time monitoring and situation detection. The real-time detection of situations allows a timely reaction and helps to reduce the damage made by harmful situations or increase the benefit of opportunities. The so called situation of interest (SOI) is described as a pattern, which represents a combination of events describing their temporal causality. Instead of just detecting SOIs, we search for possibilities to predict them. As patterns are the basis for reactive event processing, we also want to exploit them for proactive event processing. Therefore, we immerse into the world of pattern management to gain a better understanding of patterns, their structure and expressiveness.","PeriodicalId":277061,"journal":{"name":"Proceedings of the 10th ACM International Conference on Distributed and Event-based Systems","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-06-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125936122","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Real-time analysis of social networks leveraging the flink framework 利用flink框架实时分析社交网络
G. Marciani, M. Piu, M. Porretta, Matteo Nardelli, V. Cardellini
In this paper, we present a solution to the DEBS 2016 Grand Challenge that leverages Apache Flink, an open source platform for distributed stream and batch processing. We design the system architecture focusing on the exploitation of parallelism and memory efficiency so to enable an effective processing of high volume data streams on a distributed infrastructure. Our solution to the first query relies on a distributed and fine-grain approach for updating the post scores and determining partial ranks, which are then merged into a single final rank. Furthermore, changes in the final rank are identified so to update the output only if needed. The second query efficiently represents in-memory the evolving social graph and uses a customized Bron-Kerbosch algorithm to identify the largest communities active on a topic. We leverage on an in-memory caching system to keep the largest connected components which have been previously identified by the algorithm, thus saving computational time. The experimental results show that, on a portion of the dataset large half that provided for the Grand Challenge, our system can process up to 400 tuples/s with an average latency of 2.5 ms for the first query, and up to 370 tuples/s with an average latency of 2.7 ms for the second query.
在本文中,我们提出了一个利用Apache Flink(分布式流和批处理的开源平台)的DEBS 2016大挑战的解决方案。我们设计的系统架构侧重于利用并行性和内存效率,以便在分布式基础设施上有效处理大容量数据流。我们对第一个查询的解决方案依赖于一种分布式和细粒度的方法来更新帖子分数和确定部分排名,然后将其合并为单个最终排名。此外,还确定了最终排名的变化,以便仅在需要时更新输出。第二个查询有效地在内存中表示不断发展的社交图,并使用自定义的brown - kerbosch算法来识别在某个主题上活跃的最大社区。我们利用内存缓存系统来保留算法先前识别的最大连接组件,从而节省计算时间。实验结果表明,在为大挑战提供的数据集的一半大的部分上,我们的系统可以在第一次查询中处理多达400个元组/s,平均延迟为2.5 ms;在第二次查询中处理多达370个元组/s,平均延迟为2.7 ms。
{"title":"Real-time analysis of social networks leveraging the flink framework","authors":"G. Marciani, M. Piu, M. Porretta, Matteo Nardelli, V. Cardellini","doi":"10.1145/2933267.2933517","DOIUrl":"https://doi.org/10.1145/2933267.2933517","url":null,"abstract":"In this paper, we present a solution to the DEBS 2016 Grand Challenge that leverages Apache Flink, an open source platform for distributed stream and batch processing. We design the system architecture focusing on the exploitation of parallelism and memory efficiency so to enable an effective processing of high volume data streams on a distributed infrastructure. Our solution to the first query relies on a distributed and fine-grain approach for updating the post scores and determining partial ranks, which are then merged into a single final rank. Furthermore, changes in the final rank are identified so to update the output only if needed. The second query efficiently represents in-memory the evolving social graph and uses a customized Bron-Kerbosch algorithm to identify the largest communities active on a topic. We leverage on an in-memory caching system to keep the largest connected components which have been previously identified by the algorithm, thus saving computational time. The experimental results show that, on a portion of the dataset large half that provided for the Grand Challenge, our system can process up to 400 tuples/s with an average latency of 2.5 ms for the first query, and up to 370 tuples/s with an average latency of 2.7 ms for the second query.","PeriodicalId":277061,"journal":{"name":"Proceedings of the 10th ACM International Conference on Distributed and Event-based Systems","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-06-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129809512","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Monitoring top-k on real-time dynamic social-network graphs 监控top-k实时动态社交网络图
Kamalas Udomlamlert, Cosmas Krisna Adiputra, T. Hara
This paper presents our solution to 2016 DEBS Grand Challenge. We proposed our original program to efficiently calculate 2 continuous top-k queries on real-time social-network graph data. Our implementation tried to prevent processing of unaffected events by designing the algorithms to efficiently maintain the spare list of candidates of the top-k results. In addition, we improved the efficiency of the state-of-the-art algorithms to speed up the processing of the queries.
本文介绍了我们对2016年DEBS大挑战的解决方案。我们提出了我们的原始程序来有效地计算实时社交网络图数据上的2个连续top-k查询。我们的实现试图通过设计算法来有效地维护top-k结果的备用候选列表,从而防止处理未受影响的事件。此外,我们提高了最先进算法的效率,以加快查询的处理速度。
{"title":"Monitoring top-k on real-time dynamic social-network graphs","authors":"Kamalas Udomlamlert, Cosmas Krisna Adiputra, T. Hara","doi":"10.1145/2933267.2933510","DOIUrl":"https://doi.org/10.1145/2933267.2933510","url":null,"abstract":"This paper presents our solution to 2016 DEBS Grand Challenge. We proposed our original program to efficiently calculate 2 continuous top-k queries on real-time social-network graph data. Our implementation tried to prevent processing of unaffected events by designing the algorithms to efficiently maintain the spare list of candidates of the top-k results. In addition, we improved the efficiency of the state-of-the-art algorithms to speed up the processing of the queries.","PeriodicalId":277061,"journal":{"name":"Proceedings of the 10th ACM International Conference on Distributed and Event-based Systems","volume":"187 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-06-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121612352","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Distributed k-core decomposition and maintenance in large dynamic graphs 大型动态图中的分布式k核分解与维护
Sabeur Aridhi, Martin Brugnara, A. Montresor, Yannis Velegrakis
Distributed processing of large, dynamic graphs has recently received considerable attention, especially in domains such as the analytics of social networks, web graphs and spatial networks. k-core decomposition is one of the significant figures of merit that can be analyzed in graphs. Efficient algorithms to compute k-cores exist already, both in centralized and decentralized setting. Yet, these algorithms have been designed for static graphs, without significant support to deal with the addition or removal of nodes and edges. Typically, this challenge is handled by re-executing the algorithm again on the updated graph. In this work, we propose distributed k-core decomposition and maintenance algorithms for large dynamic graphs. The proposed algorithms exploit, as much as possible, the topology of the graph to compute all the k-cores and maintain them in streaming settings where edge insertions and removals happen frequently. The key idea of the maintenance strategy is that whenever the original graph is updated by the insertion/deletion of one or more edges, only a limited number of nodes need their coreness to be re-evaluated. We present an implementation of the proposed approach on top of the AKKA framework, and experimentally show the efficiency of our approach in the case of large dynamic networks.
大型动态图的分布式处理最近受到了相当大的关注,特别是在社交网络、网络图和空间网络分析等领域。k核分解是可以用图来分析的有意义的数值之一。计算k核的有效算法已经存在,无论是集中式还是分散式设置。然而,这些算法都是为静态图设计的,没有显著的支持来处理节点和边的添加或删除。通常,这个挑战是通过在更新后的图上重新执行算法来处理的。在这项工作中,我们提出了大型动态图的分布式k核分解和维护算法。所提出的算法尽可能地利用图的拓扑结构来计算所有k核,并在频繁发生边缘插入和移除的流设置中维护它们。维护策略的关键思想是,每当通过插入/删除一条或多条边来更新原始图时,只有有限数量的节点需要重新评估其核心度。我们在AKKA框架之上提出了所提出方法的实现,并通过实验证明了我们的方法在大型动态网络情况下的效率。
{"title":"Distributed k-core decomposition and maintenance in large dynamic graphs","authors":"Sabeur Aridhi, Martin Brugnara, A. Montresor, Yannis Velegrakis","doi":"10.1145/2933267.2933299","DOIUrl":"https://doi.org/10.1145/2933267.2933299","url":null,"abstract":"Distributed processing of large, dynamic graphs has recently received considerable attention, especially in domains such as the analytics of social networks, web graphs and spatial networks. k-core decomposition is one of the significant figures of merit that can be analyzed in graphs. Efficient algorithms to compute k-cores exist already, both in centralized and decentralized setting. Yet, these algorithms have been designed for static graphs, without significant support to deal with the addition or removal of nodes and edges. Typically, this challenge is handled by re-executing the algorithm again on the updated graph. In this work, we propose distributed k-core decomposition and maintenance algorithms for large dynamic graphs. The proposed algorithms exploit, as much as possible, the topology of the graph to compute all the k-cores and maintain them in streaming settings where edge insertions and removals happen frequently. The key idea of the maintenance strategy is that whenever the original graph is updated by the insertion/deletion of one or more edges, only a limited number of nodes need their coreness to be re-evaluated. We present an implementation of the proposed approach on top of the AKKA framework, and experimentally show the efficiency of our approach in the case of large dynamic networks.","PeriodicalId":277061,"journal":{"name":"Proceedings of the 10th ACM International Conference on Distributed and Event-based Systems","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-06-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115518061","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 42
Enorm: efficient window-based computation in large-scale distributed stream processing systems norm:大规模分布式流处理系统中高效的基于窗口的计算
Kasper Grud Skat Madsen, Yongluan Zhou, Li Su
Modern distributed stream processing systems (DSPS), such as Storm, typically provide a flexible programming model, where computation is specified as complicated UDFs and data is opaque to the system. While such a programming framework provides very high flexibility to the developers, it does not provide much semantic information to the system and hence it is hard to perform optimizations that has already been proved very effective in conventional stream systems. Examples include sharing computation among overlapping windows, co-partitioning operators to save communication overhead and efficient state migration during load balancing. In lieu of these challenges, we propose a new framework, which is designed to expose sufficient semantic information of the applications to enable the aforementioned effective optimizations, while on the other hand, maintaining the flexibility of Storm's original programming framework. Furthermore, we present new optimization algorithms to minimize the communication cost and state migration overhead for dynamic load balancing. We implement our framework on top of Storm and run an extensive experimental study to verify its effectiveness.
现代分布式流处理系统(DSPS),如Storm,通常提供灵活的编程模型,其中计算被指定为复杂的udf,数据对系统是不透明的。虽然这样的编程框架为开发人员提供了非常高的灵活性,但它并没有向系统提供太多的语义信息,因此很难执行在传统流系统中已经被证明非常有效的优化。例如在重叠的窗口之间共享计算,共同划分操作符以节省通信开销,以及在负载平衡期间有效的状态迁移。为了应对这些挑战,我们提出了一个新的框架,该框架旨在暴露应用程序的足够语义信息,以实现上述有效的优化,同时保持Storm原有编程框架的灵活性。此外,我们提出了新的优化算法来最小化动态负载平衡的通信开销和状态迁移开销。我们在Storm之上实现了我们的框架,并进行了广泛的实验研究来验证其有效性。
{"title":"Enorm: efficient window-based computation in large-scale distributed stream processing systems","authors":"Kasper Grud Skat Madsen, Yongluan Zhou, Li Su","doi":"10.1145/2933267.2933315","DOIUrl":"https://doi.org/10.1145/2933267.2933315","url":null,"abstract":"Modern distributed stream processing systems (DSPS), such as Storm, typically provide a flexible programming model, where computation is specified as complicated UDFs and data is opaque to the system. While such a programming framework provides very high flexibility to the developers, it does not provide much semantic information to the system and hence it is hard to perform optimizations that has already been proved very effective in conventional stream systems. Examples include sharing computation among overlapping windows, co-partitioning operators to save communication overhead and efficient state migration during load balancing. In lieu of these challenges, we propose a new framework, which is designed to expose sufficient semantic information of the applications to enable the aforementioned effective optimizations, while on the other hand, maintaining the flexibility of Storm's original programming framework. Furthermore, we present new optimization algorithms to minimize the communication cost and state migration overhead for dynamic load balancing. We implement our framework on top of Storm and run an extensive experimental study to verify its effectiveness.","PeriodicalId":277061,"journal":{"name":"Proceedings of the 10th ACM International Conference on Distributed and Event-based Systems","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-06-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124582437","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Lessons learned using a process mining approach to analyze events from distributed applications 使用流程挖掘方法分析来自分布式应用程序的事件的经验教训
Vinod Muthusamy, Aleksander Slominski, Vatche Isahagian, Rania Y. Khalaf, J. M. Reason, S. Rozsnyai
The execution of distributed applications are captured by the events generated by the individual components. However, understanding the behavior of these applications from their event logs can be a complex and error prone task, compounded by the fact that applications continuously change rendering any knowledge obsolete. We describe our experiences applying a suite of process-aware analytic tools to a number of real world scenarios, and distill our lessons learned. For example, we have seen that these tools are used iteratively, where insights gained at one stage inform the configuration decisions made at an earlier stage. As well, we have observed that data onboarding, where the raw data is cleaned and transformed, is the most critical stage in the pipeline and requires the most manual effort and domain knowledge. In particular, missing, inconsistent, and low-resolution event time stamps are recurring problems that require better solutions. The experiences and insights presented here will assist practitioners applying process analytic tools to real scenarios, and reveal to researchers some of the more pressing challenges in this space.
分布式应用程序的执行由各个组件生成的事件捕获。然而,从这些应用程序的事件日志中理解它们的行为可能是一项复杂且容易出错的任务,再加上应用程序不断变化,使得任何知识都过时了。我们描述了我们将一套流程感知分析工具应用于许多真实世界场景的经验,并总结了我们学到的经验教训。例如,我们已经看到这些工具是迭代地使用的,其中在一个阶段获得的见解通知了在早期阶段做出的配置决策。同样,我们已经观察到,在原始数据被清理和转换的地方,数据导入是管道中最关键的阶段,需要最多的手工工作和领域知识。特别是,缺少、不一致和低分辨率的事件时间戳是反复出现的问题,需要更好的解决方案。这里介绍的经验和见解将帮助实践者将过程分析工具应用于实际场景,并向研究人员揭示该领域的一些更紧迫的挑战。
{"title":"Lessons learned using a process mining approach to analyze events from distributed applications","authors":"Vinod Muthusamy, Aleksander Slominski, Vatche Isahagian, Rania Y. Khalaf, J. M. Reason, S. Rozsnyai","doi":"10.1145/2933267.2933270","DOIUrl":"https://doi.org/10.1145/2933267.2933270","url":null,"abstract":"The execution of distributed applications are captured by the events generated by the individual components. However, understanding the behavior of these applications from their event logs can be a complex and error prone task, compounded by the fact that applications continuously change rendering any knowledge obsolete. We describe our experiences applying a suite of process-aware analytic tools to a number of real world scenarios, and distill our lessons learned. For example, we have seen that these tools are used iteratively, where insights gained at one stage inform the configuration decisions made at an earlier stage. As well, we have observed that data onboarding, where the raw data is cleaned and transformed, is the most critical stage in the pipeline and requires the most manual effort and domain knowledge. In particular, missing, inconsistent, and low-resolution event time stamps are recurring problems that require better solutions. The experiences and insights presented here will assist practitioners applying process analytic tools to real scenarios, and reveal to researchers some of the more pressing challenges in this space.","PeriodicalId":277061,"journal":{"name":"Proceedings of the 10th ACM International Conference on Distributed and Event-based Systems","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-06-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116949878","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Design and development of high performance, scalable content based publish subscribe system: doctoral symposium 高性能、可扩展的基于内容的出版订阅系统的设计与开发:博士研讨会
Mrs. M. A. Shah, Walchand
In many application scenarios, content based pub/sub systems are required to provide stringent service guarantees such as reliable delivery, high performance in terms of throughput and low latency for event notification to interested subscribers. Matching algorithm play a critical role in content based pub/sub systems. The aim of our work is design and development of parallel, scalable and high performance content based publish subscribe system. We parallelize event processing using thread based and multi GPU approaches. We achieved low latency and high throughput when pub/sub is deployed on Apache Storm, a real time event processing system. Throughput gain and reduction in matching time is nearly 48% and 40% respectively in multi GPGPU approach of event processing compared to earlier work mentioned in [1].
在许多应用场景中,需要基于内容的发布/订阅系统提供严格的服务保证,例如可靠的交付、吞吐量方面的高性能以及向感兴趣的订阅者发送事件通知的低延迟。匹配算法在基于内容的发布/订阅系统中起着至关重要的作用。我们的工作目标是设计和开发并行的、可扩展的、高性能的基于内容的发布订阅系统。我们使用基于线程和多GPU的方法并行处理事件。在实时事件处理系统Apache Storm上部署pub/sub时,我们实现了低延迟和高吞吐量。与[1]中提到的早期工作相比,多GPGPU事件处理方法的吞吐量增益和匹配时间减少分别接近48%和40%。
{"title":"Design and development of high performance, scalable content based publish subscribe system: doctoral symposium","authors":"Mrs. M. A. Shah, Walchand","doi":"10.1145/2933267.2933428","DOIUrl":"https://doi.org/10.1145/2933267.2933428","url":null,"abstract":"In many application scenarios, content based pub/sub systems are required to provide stringent service guarantees such as reliable delivery, high performance in terms of throughput and low latency for event notification to interested subscribers. Matching algorithm play a critical role in content based pub/sub systems. The aim of our work is design and development of parallel, scalable and high performance content based publish subscribe system. We parallelize event processing using thread based and multi GPU approaches. We achieved low latency and high throughput when pub/sub is deployed on Apache Storm, a real time event processing system. Throughput gain and reduction in matching time is nearly 48% and 40% respectively in multi GPGPU approach of event processing compared to earlier work mentioned in [1].","PeriodicalId":277061,"journal":{"name":"Proceedings of the 10th ACM International Conference on Distributed and Event-based Systems","volume":"114 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-06-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117075842","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
OMen: overlay mending for topic-based publish/subscribe systems under churn OMen:用户流失下基于主题的发布/订阅系统的覆盖修复
Chen Chen, R. Vitenberg, H. Jacobsen
We propose, OMen, a distributed system for dynamically maintaining overlays for topic-based publish/subscribe (pub/sub) systems. In particular, OMen supports churn-resistant construction of topic-connected overlays (TCO), which organizes all nodes interested in the same topic in a directly connected dissemination sub-overlay. While aiming at pub/sub deployments in data centers, OMen internally leverages selected peer-to-peer technologies, such as T-Man as the underlying topology maintenance protocol. Existing approaches for constructing pub/sub TCOs are (i) centralized algorithms that guarantee low node degrees at the cost of prohibitive running time and (ii) decentralized protocols that are time efficient while lacking bounds on node degrees. We show both analytically and experimentally that OMen combines the best from both worlds. Namely, OMen achieves (i) low node degrees, close to centralized algorithms, and (ii) high efficiency, scalability, and load balance, comparable to decentralized protocols. Our evaluation uses both synthetic pub/sub workloads and real-world ones extracted from Facebook and Twitter. We generate churn traces with Google cluster data.
我们提出了一个分布式系统,用于动态维护基于主题的发布/订阅(pub/sub)系统的覆盖。特别是,OMen支持主题连接覆盖(TCO)的抗搅拌结构,它将对同一主题感兴趣的所有节点组织在一个直接连接的传播子覆盖中。在针对数据中心的pub/sub部署时,OMen在内部利用选定的点对点技术,如T-Man作为底层拓扑维护协议。现有的构建pub/sub tco的方法是(i)以高昂的运行时间为代价保证低节点度的集中式算法和(ii)时间效率高但缺乏节点度限制的分散协议。我们通过分析和实验证明,OMen结合了这两个世界的优点。也就是说,OMen实现了(1)低节点度,接近集中式算法;(2)高效率、可扩展性和负载均衡,与去中心化协议相当。我们的评估使用了合成的发布/订阅工作负载和从Facebook和Twitter中提取的真实工作负载。我们使用Google集群数据生成流失痕迹。
{"title":"OMen: overlay mending for topic-based publish/subscribe systems under churn","authors":"Chen Chen, R. Vitenberg, H. Jacobsen","doi":"10.1145/2933267.2933305","DOIUrl":"https://doi.org/10.1145/2933267.2933305","url":null,"abstract":"We propose, OMen, a distributed system for dynamically maintaining overlays for topic-based publish/subscribe (pub/sub) systems. In particular, OMen supports churn-resistant construction of topic-connected overlays (TCO), which organizes all nodes interested in the same topic in a directly connected dissemination sub-overlay. While aiming at pub/sub deployments in data centers, OMen internally leverages selected peer-to-peer technologies, such as T-Man as the underlying topology maintenance protocol. Existing approaches for constructing pub/sub TCOs are (i) centralized algorithms that guarantee low node degrees at the cost of prohibitive running time and (ii) decentralized protocols that are time efficient while lacking bounds on node degrees. We show both analytically and experimentally that OMen combines the best from both worlds. Namely, OMen achieves (i) low node degrees, close to centralized algorithms, and (ii) high efficiency, scalability, and load balance, comparable to decentralized protocols. Our evaluation uses both synthetic pub/sub workloads and real-world ones extracted from Facebook and Twitter. We generate churn traces with Google cluster data.","PeriodicalId":277061,"journal":{"name":"Proceedings of the 10th ACM International Conference on Distributed and Event-based Systems","volume":"61 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-06-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123464645","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Data-centric publish/subscribe routing middleware for realizing proactive overlay software-defined networking 以数据为中心的发布/订阅路由中间件,用于实现主动覆盖软件定义网络
Akram Hakiri, A. Gokhale
Software Defined Networking (SDN) has emerged as an attractive solution to allow cloud-to-cloud interconnection and federation. SDN technologies, such as OpenFlow, use both reactive hop-by-hop and proactive approaches to program the switches. The reactive strategy incurs substantial scalability problems for large networks due to the hop-by-hop behavior while the proactive approach is hard to implement in practice due to the need to forecast all possible forwarding rules ahead-of-time. An attractive and more realistic alternative is the proactive overlay SDN approach, however, many challenges must first be overcome to realize it. Existing techniques to program the switches use low-level programming abstractions, which are error-prone and cannot scale. Middleware-based solutions, e.g., using XMPP, are stateful and hence also incur substantial scalability issues. Although content-based publish/subscribe (pub/sub) solutions have been used in the past for SDN, they rely on brokers, which is problematic and incurs unnecessary additional infrastructure elements that pollute the SDN architecture. To address these issues, this paper demonstrates how the strengths of the data-centric, broker-less pub/sub paradigm can be exploited to realize proactive overlay SDN for inter cloud domain federation. To that end, we first present the design rationale and architecture of our solution called POSEIDON (Proactive brOkerless SubscribEr Interest-Defined Overlay Networking). Second, we present the messaging protocol between the controller and switches. Finally, we present results of evaluating POSEIDON and illustrate how it improves data delivery and provides high performance at the network-level in proactive overlay SDN.
软件定义网络(SDN)已经成为一种有吸引力的解决方案,允许云到云的互连和联合。SDN技术,如OpenFlow,使用被动的逐跳和主动的方法来编程交换机。在大型网络中,被动策略由于其逐跳行为而导致了严重的可扩展性问题,而主动策略由于需要提前预测所有可能的转发规则而难以在实践中实现。一种有吸引力且更现实的替代方案是主动覆盖SDN方法,然而,要实现它必须首先克服许多挑战。现有的交换机编程技术使用低级编程抽象,容易出错且无法扩展。基于中间件的解决方案(例如,使用XMPP)是有状态的,因此也会导致大量的可伸缩性问题。尽管基于内容的发布/订阅(pub/sub)解决方案在过去已经用于SDN,但它们依赖于代理,这是有问题的,并且会产生不必要的额外基础设施元素,从而污染SDN体系结构。为了解决这些问题,本文展示了如何利用以数据为中心、无代理的发布/订阅范式的优势来实现云域间联合的主动覆盖SDN。为此,我们首先介绍了我们的解决方案POSEIDON(主动无代理订户兴趣定义覆盖网络)的设计原理和体系结构。其次,给出了控制器与交换机之间的消息传递协议。最后,我们介绍了评估POSEIDON的结果,并说明了它如何在主动覆盖SDN中改善数据传输并提供网络级的高性能。
{"title":"Data-centric publish/subscribe routing middleware for realizing proactive overlay software-defined networking","authors":"Akram Hakiri, A. Gokhale","doi":"10.1145/2933267.2933314","DOIUrl":"https://doi.org/10.1145/2933267.2933314","url":null,"abstract":"Software Defined Networking (SDN) has emerged as an attractive solution to allow cloud-to-cloud interconnection and federation. SDN technologies, such as OpenFlow, use both reactive hop-by-hop and proactive approaches to program the switches. The reactive strategy incurs substantial scalability problems for large networks due to the hop-by-hop behavior while the proactive approach is hard to implement in practice due to the need to forecast all possible forwarding rules ahead-of-time. An attractive and more realistic alternative is the proactive overlay SDN approach, however, many challenges must first be overcome to realize it. Existing techniques to program the switches use low-level programming abstractions, which are error-prone and cannot scale. Middleware-based solutions, e.g., using XMPP, are stateful and hence also incur substantial scalability issues. Although content-based publish/subscribe (pub/sub) solutions have been used in the past for SDN, they rely on brokers, which is problematic and incurs unnecessary additional infrastructure elements that pollute the SDN architecture. To address these issues, this paper demonstrates how the strengths of the data-centric, broker-less pub/sub paradigm can be exploited to realize proactive overlay SDN for inter cloud domain federation. To that end, we first present the design rationale and architecture of our solution called POSEIDON (Proactive brOkerless SubscribEr Interest-Defined Overlay Networking). Second, we present the messaging protocol between the controller and switches. Finally, we present results of evaluating POSEIDON and illustrate how it improves data delivery and provides high performance at the network-level in proactive overlay SDN.","PeriodicalId":277061,"journal":{"name":"Proceedings of the 10th ACM International Conference on Distributed and Event-based Systems","volume":"321 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-06-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133040985","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Enabling semantic access to static and streaming distributed data with optique: demo 使用optique: demo实现对静态和流分布式数据的语义访问
E. Kharlamov, S. Brandt, M. Giese, Ernesto Jiménez-Ruiz, Y. Kotidis, S. Lamparter, T. Mailis, C. Neuenstadt, Ö. Özçep, C. Pinkel, A. Soylu, C. Svingos, D. Zheleznyakov, Ian Horrocks, Y. Ioannidis, R. Möller, A. Waaler
Real-time processing of data coming from multiple heterogeneous data streams and static databases is a typical task in many industrial scenarios such as diagnostics of large machines. A complex diagnostic task may require a collection of up to hundreds of queries over such data. Although many of these queries retrieve data of the same kind, such as temperature measurements, they access structurally different data sources. In this work, we show how Semantic Technologies implemented in our system Optique can simplify such complex diagnostics by providing an abstraction layer---ontology---that integrates heterogeneous data. In a nutshell, Optique allows complex diagnostic tasks to be expressed with just a few high-level semantic queries, which can be easily formulated with our visual query formulation system. Optique can then automatically enrich these queries, translate them into a large collection of low-level data queries, and finally optimise and efficiently execute the collection in a heavily distributed environment.
实时处理来自多个异构数据流和静态数据库的数据是许多工业场景中的典型任务,例如大型机器的诊断。复杂的诊断任务可能需要对此类数据进行多达数百次查询的集合。尽管这些查询中有许多检索相同类型的数据,例如温度测量值,但它们访问的数据源在结构上是不同的。在这项工作中,我们展示了在我们的系统Optique中实现的语义技术如何通过提供集成异构数据的抽象层——本体——来简化这种复杂的诊断。简而言之,Optique允许用几个高级语义查询来表达复杂的诊断任务,这些查询可以很容易地用我们的视觉查询公式系统来制定。Optique可以自动丰富这些查询,将它们转换为大量低级数据查询,最后在高度分布式的环境中优化并有效地执行这些查询。
{"title":"Enabling semantic access to static and streaming distributed data with optique: demo","authors":"E. Kharlamov, S. Brandt, M. Giese, Ernesto Jiménez-Ruiz, Y. Kotidis, S. Lamparter, T. Mailis, C. Neuenstadt, Ö. Özçep, C. Pinkel, A. Soylu, C. Svingos, D. Zheleznyakov, Ian Horrocks, Y. Ioannidis, R. Möller, A. Waaler","doi":"10.1145/2933267.2933290","DOIUrl":"https://doi.org/10.1145/2933267.2933290","url":null,"abstract":"Real-time processing of data coming from multiple heterogeneous data streams and static databases is a typical task in many industrial scenarios such as diagnostics of large machines. A complex diagnostic task may require a collection of up to hundreds of queries over such data. Although many of these queries retrieve data of the same kind, such as temperature measurements, they access structurally different data sources. In this work, we show how Semantic Technologies implemented in our system Optique can simplify such complex diagnostics by providing an abstraction layer---ontology---that integrates heterogeneous data. In a nutshell, Optique allows complex diagnostic tasks to be expressed with just a few high-level semantic queries, which can be easily formulated with our visual query formulation system. Optique can then automatically enrich these queries, translate them into a large collection of low-level data queries, and finally optimise and efficiently execute the collection in a heavily distributed environment.","PeriodicalId":277061,"journal":{"name":"Proceedings of the 10th ACM International Conference on Distributed and Event-based Systems","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-06-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125390818","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
期刊
Proceedings of the 10th ACM International Conference on Distributed and Event-based Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1