首页 > 最新文献

2021 IEEE 27th International Conference on Parallel and Distributed Systems (ICPADS)最新文献

英文 中文
BRNN-GAN: Generative Adversarial Networks with Bi-directional Recurrent Neural Networks for Multivariate Time Series Imputation 基于双向递归神经网络的多元时间序列插值生成对抗网络
Pub Date : 2021-12-01 DOI: 10.1109/ICPADS53394.2021.00033
Zejun Wu, Chao Ma, Xiaochuan Shi, Libing Wu, Dian Zhang, Yutian Tang, M. Stojmenovic
Missing values appearing in multivariate time series often prevent further and in-depth analysis in real-world applications. To handle those missing values, advanced multivariate time series imputation methods are expected to (1) consider bi-directional temporal correlations, (2) model cross-variable correlations, and (3) approximate original data's distribution. However, most of existing approaches are not able to meet all the three above-mentioned requirements. Drawing on advances in machine learning, we propose BRNN-GAN, a generative adversarial network with bi-directional RNN cells. The BRNN cell is designed to model bi-directional temporal and cross-variable correlations, and the GAN architecture is employed to learn original data's distribution. By conducting comprehensive experiments on two public datasets, the experimental results show that our proposed BRNN-GAN outperforms all the baselines in terms of achieving the lowest Mean Absolute Error (MAE).
在实际应用中,多变量时间序列中出现的缺失值通常会阻碍进一步深入的分析。为了处理这些缺失值,先进的多变量时间序列imputation方法需要:(1)考虑双向时间相关性,(2)建立交叉变量相关性模型,(3)近似原始数据的分布。然而,现有的大多数方法都不能满足上述三个要求。利用机器学习的进步,我们提出了BRNN-GAN,一种双向RNN细胞的生成对抗网络。BRNN单元被设计用来模拟双向时间和交叉变量的相关性,GAN架构被用来学习原始数据的分布。通过在两个公共数据集上进行综合实验,实验结果表明,我们提出的BRNN-GAN在实现最低平均绝对误差(MAE)方面优于所有基线。
{"title":"BRNN-GAN: Generative Adversarial Networks with Bi-directional Recurrent Neural Networks for Multivariate Time Series Imputation","authors":"Zejun Wu, Chao Ma, Xiaochuan Shi, Libing Wu, Dian Zhang, Yutian Tang, M. Stojmenovic","doi":"10.1109/ICPADS53394.2021.00033","DOIUrl":"https://doi.org/10.1109/ICPADS53394.2021.00033","url":null,"abstract":"Missing values appearing in multivariate time series often prevent further and in-depth analysis in real-world applications. To handle those missing values, advanced multivariate time series imputation methods are expected to (1) consider bi-directional temporal correlations, (2) model cross-variable correlations, and (3) approximate original data's distribution. However, most of existing approaches are not able to meet all the three above-mentioned requirements. Drawing on advances in machine learning, we propose BRNN-GAN, a generative adversarial network with bi-directional RNN cells. The BRNN cell is designed to model bi-directional temporal and cross-variable correlations, and the GAN architecture is employed to learn original data's distribution. By conducting comprehensive experiments on two public datasets, the experimental results show that our proposed BRNN-GAN outperforms all the baselines in terms of achieving the lowest Mean Absolute Error (MAE).","PeriodicalId":309508,"journal":{"name":"2021 IEEE 27th International Conference on Parallel and Distributed Systems (ICPADS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129647240","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Delica: Decentralized Lightweight Collective Attestation for Disruptive IoT Networks Delica:颠覆性物联网网络的去中心化轻量级集体认证
Pub Date : 2021-12-01 DOI: 10.1109/ICPADS53394.2021.00051
Ziyu Wang, Cong Sun, Qingsong Yao, Duo Ding, Jianfeng Ma
The recent advance of the Internet of Things and autonomous systems brings massive security threats to the network of low-end embedded devices. Remote attestation is a hardware-assisted technique to verify the integrity and trustworthiness of software on remote devices. The recently proposed collective remote attestations have focused on attesting to the highly dynamic and disruptive device networks. However, they are generally inefficient due to the homogeneous node setting for the robustness of attestation reports aggregation. In this work, we propose Delica, an efficient and robust collective attestation framework for dynamic and disruptive networks. We differentiate the role of provers and aggregators to limit the redundant communications and attestation evidence aggregations for efficiency. Delica is capable of mitigating DoS attacks and detecting physical and black-hole attacks. The experimental results and analysis show that Delica can greatly reduce the per-node computational cost and reduce the network attestation cost by over 75% compared with the state-of-the-art approaches on disruptive networks.
近年来,物联网和自主系统的发展给低端嵌入式设备网络带来了巨大的安全威胁。远程认证是一种硬件辅助技术,用于验证远程设备上软件的完整性和可信度。最近提出的集体远程认证侧重于对高度动态和破坏性的设备网络进行认证。然而,由于为保证认证报告聚合的健壮性而采用同构节点设置,它们通常效率低下。在这项工作中,我们提出了Delica,这是一个针对动态和破坏性网络的高效且强大的集体认证框架。我们区分证明者和聚合者的角色,以限制冗余通信和证明证据聚合,提高效率。Delica能够减轻DoS攻击,并检测物理和黑洞攻击。实验结果和分析表明,与目前最先进的破坏性网络方法相比,Delica可以大大降低每个节点的计算成本,将网络认证成本降低75%以上。
{"title":"Delica: Decentralized Lightweight Collective Attestation for Disruptive IoT Networks","authors":"Ziyu Wang, Cong Sun, Qingsong Yao, Duo Ding, Jianfeng Ma","doi":"10.1109/ICPADS53394.2021.00051","DOIUrl":"https://doi.org/10.1109/ICPADS53394.2021.00051","url":null,"abstract":"The recent advance of the Internet of Things and autonomous systems brings massive security threats to the network of low-end embedded devices. Remote attestation is a hardware-assisted technique to verify the integrity and trustworthiness of software on remote devices. The recently proposed collective remote attestations have focused on attesting to the highly dynamic and disruptive device networks. However, they are generally inefficient due to the homogeneous node setting for the robustness of attestation reports aggregation. In this work, we propose Delica, an efficient and robust collective attestation framework for dynamic and disruptive networks. We differentiate the role of provers and aggregators to limit the redundant communications and attestation evidence aggregations for efficiency. Delica is capable of mitigating DoS attacks and detecting physical and black-hole attacks. The experimental results and analysis show that Delica can greatly reduce the per-node computational cost and reduce the network attestation cost by over 75% compared with the state-of-the-art approaches on disruptive networks.","PeriodicalId":309508,"journal":{"name":"2021 IEEE 27th International Conference on Parallel and Distributed Systems (ICPADS)","volume":"462 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129567338","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Highly Scalable Parallel Checksums 高度可扩展的并行校验和
Pub Date : 2021-12-01 DOI: 10.1109/ICPADS53394.2021.00107
Christian Siebert
Checksums are used to detect errors that might occur while storing or communicating data. Checking the integrity of data is well-established, but only for smaller data sets. Contrary, supercomputers have to deal with huge amounts of data, which introduces failures that may remain undetected. Therefore, additional protection becomes a necessity at large scale. However, checking the integrity of larger data sets, especially in case of distributed data, clearly requires parallel approaches. We show how popular checksums, such as CRC-32 or Adler-32, can be parallelized efficiently. This also disproves a widespread belief that parallelizing aforementioned checksums, especially in a scalable way, is not possible. The mathematical properties behind these checksums enable a method to combine partial checksums such that its result corresponds to the checksum of the concatenated partial data. Our parallel checksum algorithm utilizes this combination idea in a scalable hierarchical reduction scheme to combine the partial checksums from an arbitrary number of processing elements. Although this reduction scheme can be implemented manually using most parallel programming interfaces, we use the Message Passing Interface, which supports such a functionality directly via non-commutative user-defined reduction operations. In conjunction with the efficient checksum capabilities of the zlib library, our algorithm can not only be implemented conveniently and in a portable way, but also very efficiently. Additional shared-memory parallelization within compute nodes completes our hybrid parallel checksum solutions, which show a high scalability of up to 524,288 threads. At this scale, computing the checksums of 240 TiB data took only 3.4 seconds for CRC-32 and 2.6 seconds for Adler-32. Finally, we discuss the APES application as a representative of dynamic supercomputer applications. Thanks to our scalable checksum algorithm, even such applications are now able to detect many errors within their distributed data sets.
校验和用于检测在存储或通信数据时可能发生的错误。检查数据的完整性已经建立,但只适用于较小的数据集。相反,超级计算机必须处理大量数据,这可能会导致未被发现的故障。因此,在大范围内,额外的保护是必要的。然而,检查大型数据集的完整性,特别是在分布式数据的情况下,显然需要并行方法。我们展示了流行的校验和(如CRC-32或Adler-32)如何有效地并行化。这也反驳了一个普遍的观点,即并行化前面提到的校验和,特别是以可伸缩的方式,是不可能的。这些校验和背后的数学属性使方法能够组合部分校验和,使其结果与连接的部分数据的校验和相对应。我们的并行校验和算法在可伸缩的分层约简方案中利用这种组合思想来组合来自任意数量的处理元素的部分校验和。虽然这种简化方案可以使用大多数并行编程接口手动实现,但我们使用消息传递接口,它通过非交换的用户定义简化操作直接支持这种功能。结合zlib库的有效校验和功能,我们的算法不仅可以方便地以可移植的方式实现,而且非常高效。计算节点内的额外共享内存并行化完成了我们的混合并行校验和解决方案,它显示了高达524,288个线程的高可伸缩性。在这个规模下,计算240 TiB数据的校验和对于CRC-32只需要3.4秒,对于Adler-32只需要2.6秒。最后,我们讨论了作为动态超级计算机应用代表的APES应用。由于我们的可扩展校验和算法,即使是这样的应用程序现在也能够检测到其分布式数据集中的许多错误。
{"title":"Highly Scalable Parallel Checksums","authors":"Christian Siebert","doi":"10.1109/ICPADS53394.2021.00107","DOIUrl":"https://doi.org/10.1109/ICPADS53394.2021.00107","url":null,"abstract":"Checksums are used to detect errors that might occur while storing or communicating data. Checking the integrity of data is well-established, but only for smaller data sets. Contrary, supercomputers have to deal with huge amounts of data, which introduces failures that may remain undetected. Therefore, additional protection becomes a necessity at large scale. However, checking the integrity of larger data sets, especially in case of distributed data, clearly requires parallel approaches. We show how popular checksums, such as CRC-32 or Adler-32, can be parallelized efficiently. This also disproves a widespread belief that parallelizing aforementioned checksums, especially in a scalable way, is not possible. The mathematical properties behind these checksums enable a method to combine partial checksums such that its result corresponds to the checksum of the concatenated partial data. Our parallel checksum algorithm utilizes this combination idea in a scalable hierarchical reduction scheme to combine the partial checksums from an arbitrary number of processing elements. Although this reduction scheme can be implemented manually using most parallel programming interfaces, we use the Message Passing Interface, which supports such a functionality directly via non-commutative user-defined reduction operations. In conjunction with the efficient checksum capabilities of the zlib library, our algorithm can not only be implemented conveniently and in a portable way, but also very efficiently. Additional shared-memory parallelization within compute nodes completes our hybrid parallel checksum solutions, which show a high scalability of up to 524,288 threads. At this scale, computing the checksums of 240 TiB data took only 3.4 seconds for CRC-32 and 2.6 seconds for Adler-32. Finally, we discuss the APES application as a representative of dynamic supercomputer applications. Thanks to our scalable checksum algorithm, even such applications are now able to detect many errors within their distributed data sets.","PeriodicalId":309508,"journal":{"name":"2021 IEEE 27th International Conference on Parallel and Distributed Systems (ICPADS)","volume":"75 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124986170","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving Restore Performance of Deduplication Systems via a Greedy Rewriting Scheme 利用贪婪重写方案提高重删系统的恢复性能
Pub Date : 2021-12-01 DOI: 10.1109/ICPADS53394.2021.00042
Lifang Lin, Yuhui Deng, Yi Zhou
Data deduplication has been widely used to improve storage space utilization, however, it is baffled by data fragmen-tation: logically consecutive chunks physically scattered across various containers. Many rewriting schemes, rewriting fragment-ed duplicate chunks into new containers, attempt to alleviate the restore performance degradation caused by fragmentation. Unfortunately, these schemes rely on a fixed threshold and fail to choose the appropriate set of old containers for rewriting, which leads to substantial redundant chunks existing in the retrieved containers when restoring backups. To address this issue, we propose a flexible threshold rewriting scheme to improve restore performance while maintaining high backup performance. We define an effectiveness metric - valid container reference counts (VCRC) - that facilitates identifying the appropriate containers for rewriting. We design a greedy-algorithm-based algorithm called F-greedy that dynamically adjusts the threshold according to the distribution of containers' VCRC, aiming to rewrite low-VCRC containers. We quantitatively evaluate F-greedy on three real-world backup datasets in terms of restore performance, backup performance, and storage overhead. The empirical results show that compared with two state-of-the-art schemes (Capping and SMR), our scheme improves the restore speed of the exiting algorithms by 1.3x - 2.4x while achieving similar backup performance.
重复数据删除已被广泛用于提高存储空间利用率,然而,它受到数据碎片的困扰:逻辑上连续的块物理上分散在不同的容器中。许多重写方案(将碎片化的重复块重写到新的容器中)都试图减轻碎片造成的恢复性能下降。不幸的是,这些方案依赖于一个固定的阈值,并且不能选择适当的旧容器集进行重写,这导致在恢复备份时检索到的容器中存在大量冗余块。为了解决这个问题,我们提出了一种灵活的阈值重写方案,以提高恢复性能,同时保持较高的备份性能。我们定义了一个有效性度量——有效容器引用计数(VCRC)——它有助于识别要重写的适当容器。我们设计了一种基于贪婪算法的F-greedy算法,根据容器的VCRC分布动态调整阈值,旨在重写低VCRC的容器。我们从恢复性能、备份性能和存储开销方面定量地评估了三个真实备份数据集上的F-greedy。实证结果表明,与两种最先进的方案(Capping和SMR)相比,我们的方案在达到相似的备份性能的同时,将现有算法的恢复速度提高了1.3 - 2.4倍。
{"title":"Improving Restore Performance of Deduplication Systems via a Greedy Rewriting Scheme","authors":"Lifang Lin, Yuhui Deng, Yi Zhou","doi":"10.1109/ICPADS53394.2021.00042","DOIUrl":"https://doi.org/10.1109/ICPADS53394.2021.00042","url":null,"abstract":"Data deduplication has been widely used to improve storage space utilization, however, it is baffled by data fragmen-tation: logically consecutive chunks physically scattered across various containers. Many rewriting schemes, rewriting fragment-ed duplicate chunks into new containers, attempt to alleviate the restore performance degradation caused by fragmentation. Unfortunately, these schemes rely on a fixed threshold and fail to choose the appropriate set of old containers for rewriting, which leads to substantial redundant chunks existing in the retrieved containers when restoring backups. To address this issue, we propose a flexible threshold rewriting scheme to improve restore performance while maintaining high backup performance. We define an effectiveness metric - valid container reference counts (VCRC) - that facilitates identifying the appropriate containers for rewriting. We design a greedy-algorithm-based algorithm called F-greedy that dynamically adjusts the threshold according to the distribution of containers' VCRC, aiming to rewrite low-VCRC containers. We quantitatively evaluate F-greedy on three real-world backup datasets in terms of restore performance, backup performance, and storage overhead. The empirical results show that compared with two state-of-the-art schemes (Capping and SMR), our scheme improves the restore speed of the exiting algorithms by 1.3x - 2.4x while achieving similar backup performance.","PeriodicalId":309508,"journal":{"name":"2021 IEEE 27th International Conference on Parallel and Distributed Systems (ICPADS)","volume":"204 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122839659","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Sensitivity loss training based implicit feedback 基于隐式反馈的灵敏度损失训练
Pub Date : 2021-12-01 DOI: 10.1109/ICPADS53394.2021.00036
Kunyu Li, Nan Wang, Xinyu Liu
In recommender systems, due to the lack of explicit feedback features, datasets with implicit feedback are always accustomed to train all samples without separating them during model training, without considering the non-consistency of samples. This leads to a significant decrease in sample utilization and creates challenges for model training. Also, little work has been done to explore the intrinsic laws implied in the implicit feedback dataset and how to effectively train the implicit feedback data. In this paper, we first summarize the variation pattern of loss with model training for different rating samples in the explicit feedback dataset, and find that model training is highly sensitive to the ratings. Second, we design an adaptive hierarchical training function with dynamic thresholds that can effectively distinguish different rating samples in the dataset, thus optimizing the implicit feedback dataset into an explicit feedback dataset to some extent. Finally, to better learn samples with different ratings, we also propose an adaptive hierarchical training strategy to obtain better training results in the implicit feedback dataset. Extensive experiments on three datasets show that our approach achieves excellent performance and greatly improves the performance of the model.
在推荐系统中,由于缺乏显式反馈特征,具有隐式反馈的数据集在模型训练时总是习惯于训练所有样本而不进行分离,而不考虑样本的不一致性。这导致样本利用率显著降低,并为模型训练带来挑战。此外,关于隐式反馈数据集隐含的内在规律以及如何有效训练隐式反馈数据的研究也很少。本文首先总结了显式反馈数据集中不同评级样本的损失随模型训练的变化规律,发现模型训练对评级高度敏感。其次,设计具有动态阈值的自适应分层训练函数,有效区分数据集中不同评级样本,从而在一定程度上将隐式反馈数据集优化为显式反馈数据集。最后,为了更好地学习不同评级的样本,我们还提出了一种自适应分层训练策略,以在隐式反馈数据集中获得更好的训练结果。在三个数据集上的大量实验表明,我们的方法取得了优异的性能,大大提高了模型的性能。
{"title":"Sensitivity loss training based implicit feedback","authors":"Kunyu Li, Nan Wang, Xinyu Liu","doi":"10.1109/ICPADS53394.2021.00036","DOIUrl":"https://doi.org/10.1109/ICPADS53394.2021.00036","url":null,"abstract":"In recommender systems, due to the lack of explicit feedback features, datasets with implicit feedback are always accustomed to train all samples without separating them during model training, without considering the non-consistency of samples. This leads to a significant decrease in sample utilization and creates challenges for model training. Also, little work has been done to explore the intrinsic laws implied in the implicit feedback dataset and how to effectively train the implicit feedback data. In this paper, we first summarize the variation pattern of loss with model training for different rating samples in the explicit feedback dataset, and find that model training is highly sensitive to the ratings. Second, we design an adaptive hierarchical training function with dynamic thresholds that can effectively distinguish different rating samples in the dataset, thus optimizing the implicit feedback dataset into an explicit feedback dataset to some extent. Finally, to better learn samples with different ratings, we also propose an adaptive hierarchical training strategy to obtain better training results in the implicit feedback dataset. Extensive experiments on three datasets show that our approach achieves excellent performance and greatly improves the performance of the model.","PeriodicalId":309508,"journal":{"name":"2021 IEEE 27th International Conference on Parallel and Distributed Systems (ICPADS)","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122014212","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Boosting Byzantine Protocols in Large Sparse Networks with High System Assumption Coverage 高系统假设覆盖率的大型稀疏网络中拜占庭协议的增强
Pub Date : 2021-12-01 DOI: 10.1109/ICPADS53394.2021.00097
Shaolin Yu, Jihong Zhu, Jiali Yang, Yulong Zhan
To improve the overall efficiency and reliability of Byzantine protocols in large sparse networks, we propose a new system assumption for developing multi-scale fault-tolerant systems, with which several kinds of multi-scale Byzantine protocols are developed in large sparse networks with high system assumption coverage. By extending the traditional Byzantine adversary to the multi-scale adversaries, it is shown that efficient deterministic Byzantine broadcast and Byzantine agreement can be built in logarithmic-degree networks. Meanwhile, it is shown that the multi-scale adversary can make a finer trade-off between the system assumption coverage and the overall efficiency of the Byzantine protocols, especially when a small portion of the low-layer small-scale protocols are allowed to fail arbitrarily. With this, efficient Byzantine protocols can be built in large sparse networks with high system reliability.
为了提高拜占庭协议在大型稀疏网络中的整体效率和可靠性,提出了开发多尺度容错系统的新系统假设,并在此基础上开发了几种具有高系统假设覆盖率的大型稀疏网络中的多尺度拜占庭协议。通过将传统的拜占庭对手扩展到多尺度对手,证明了在对数度网络中可以建立有效的确定性拜占庭广播和拜占庭协议。同时,研究表明,多尺度攻击者可以在系统假设覆盖率和拜占庭协议的整体效率之间做出更好的权衡,特别是当允许一小部分低层小规模协议任意失效时。利用这种方法,可以在大型稀疏网络中构建高效的拜占庭协议,并具有较高的系统可靠性。
{"title":"Boosting Byzantine Protocols in Large Sparse Networks with High System Assumption Coverage","authors":"Shaolin Yu, Jihong Zhu, Jiali Yang, Yulong Zhan","doi":"10.1109/ICPADS53394.2021.00097","DOIUrl":"https://doi.org/10.1109/ICPADS53394.2021.00097","url":null,"abstract":"To improve the overall efficiency and reliability of Byzantine protocols in large sparse networks, we propose a new system assumption for developing multi-scale fault-tolerant systems, with which several kinds of multi-scale Byzantine protocols are developed in large sparse networks with high system assumption coverage. By extending the traditional Byzantine adversary to the multi-scale adversaries, it is shown that efficient deterministic Byzantine broadcast and Byzantine agreement can be built in logarithmic-degree networks. Meanwhile, it is shown that the multi-scale adversary can make a finer trade-off between the system assumption coverage and the overall efficiency of the Byzantine protocols, especially when a small portion of the low-layer small-scale protocols are allowed to fail arbitrarily. With this, efficient Byzantine protocols can be built in large sparse networks with high system reliability.","PeriodicalId":309508,"journal":{"name":"2021 IEEE 27th International Conference on Parallel and Distributed Systems (ICPADS)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121059376","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
RTPoW: A Proof-of-Work Consensus Scheme with Real-Time Difficulty Adjustment Algorithm RTPoW:一种具有实时难度调整算法的工作量证明共识方案
Pub Date : 2021-12-01 DOI: 10.1109/ICPADS53394.2021.00035
Weijia Feng, Zhenfu Cao, Jiachen Shen, Xiaolei Dong
Bitcoin, the first decentralized cryptocurrency system, uses a simple but effective difficulty adjustment algorithm to stabilize its average time of the block creation at 10 minutes. Over time, the volatility of the Bitcoin price has become higher and higher, and it causes the total hashrate (the hash power of the entire network) constantly fluctuating. Both facts and our experimental results prove that Bitcoin's difficulty adjustment algorithm cannot respond in time while the total hashrate is constantly fluctuating. Hence, we propose a consensus protocol with a real-time difficulty adjustment algorithm, RTPoW. RTPoW allows the blockchain to adjust the difficulty target of each block by predicting the real-time total hashrate, so the block time can remain stable even if the total hashrate is wildly fluctuating. To evaluate the effect of RTPoW, we implemented a simulator of an experimental environment and tested our algorithm. The results obtained have confirmed its effectiveness and stability.
比特币是第一个去中心化的加密货币系统,它使用了一种简单而有效的难度调整算法,将区块创建的平均时间稳定在10分钟。随着时间的推移,比特币价格的波动性越来越大,导致总哈希率(整个网络的哈希算力)不断波动。事实和我们的实验结果都证明,当总哈希值不断波动时,比特币的难度调整算法无法及时响应。因此,我们提出了一种带有实时难度调整算法RTPoW的共识协议。RTPoW允许区块链通过预测实时总哈希率来调整每个区块的难度目标,因此即使总哈希率剧烈波动,区块时间也可以保持稳定。为了评估RTPoW的效果,我们实现了一个实验环境的模拟器并测试了我们的算法。实验结果证实了该方法的有效性和稳定性。
{"title":"RTPoW: A Proof-of-Work Consensus Scheme with Real-Time Difficulty Adjustment Algorithm","authors":"Weijia Feng, Zhenfu Cao, Jiachen Shen, Xiaolei Dong","doi":"10.1109/ICPADS53394.2021.00035","DOIUrl":"https://doi.org/10.1109/ICPADS53394.2021.00035","url":null,"abstract":"Bitcoin, the first decentralized cryptocurrency system, uses a simple but effective difficulty adjustment algorithm to stabilize its average time of the block creation at 10 minutes. Over time, the volatility of the Bitcoin price has become higher and higher, and it causes the total hashrate (the hash power of the entire network) constantly fluctuating. Both facts and our experimental results prove that Bitcoin's difficulty adjustment algorithm cannot respond in time while the total hashrate is constantly fluctuating. Hence, we propose a consensus protocol with a real-time difficulty adjustment algorithm, RTPoW. RTPoW allows the blockchain to adjust the difficulty target of each block by predicting the real-time total hashrate, so the block time can remain stable even if the total hashrate is wildly fluctuating. To evaluate the effect of RTPoW, we implemented a simulator of an experimental environment and tested our algorithm. The results obtained have confirmed its effectiveness and stability.","PeriodicalId":309508,"journal":{"name":"2021 IEEE 27th International Conference on Parallel and Distributed Systems (ICPADS)","volume":"95 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124690969","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FerryLink: Combating Link Degradation for Practical LPWAN Deployments FerryLink:为实际的LPWAN部署对抗链路退化
Pub Date : 2021-12-01 DOI: 10.1109/ICPADS53394.2021.00077
Jing Yang, Zhenqiang Xu, Jiliang Wang
Low-Power Wide-Area Networks (LPWANs) have been shown as a promising technique to provide long-range low-power communication for large-scale IoT devices. In this paper, however, we show the poor performance of LoRa network due to its link diversity in macro- and micro- scope through one-month measurements in an area of $2.2 kmtimes 1.5 km$. We present FerryLink, which exploits such link diversity and leverages peer nodes to ferry data of weak links, to combat performance degradation. Traditional arts (e.g., building multi-hop networks) are inefficient or too heavyweight for the current star-topology-based LoRa network. FerryLink thus proposes a novel ferry mechanism combining RSSI sampling and Channel Activity detection(CAD) to suit multiple orthogonal transmission parameters of LoRa. To reduce energy overhead, FerryLink leverages convention windows for coarse-grained transmission synchronization between two coupled nodes. Finally, FerryLink utilizes the orthogonality of uplink and downlink signals to avoid data redundancy due to the ferry mechanism, maintaining comparable capacity with original LPWANs. We build FerryLink on top of LoRaWANwith commercial off-the-shelf hardware. The extensive evaluation results show that FerryLink effectively improves the packet delivery rate (PDR) of LoRa nodes (to over 95%), achieves 2x less energy overhead, and increases communication range by 50% compared with the original LoRaWAN.
低功耗广域网(lpwan)已被证明是为大规模物联网设备提供远程低功耗通信的一种有前途的技术。然而,在本文中,我们通过在2.2 km乘以1.5 km$的区域内进行为期一个月的测量,表明LoRa网络由于其宏观和微观的链路多样性而性能不佳。我们提出了FerryLink,它利用这种链路多样性并利用对等节点来传送弱链路的数据,以对抗性能下降。传统技术(例如,构建多跳网络)对于当前基于星型拓扑的LoRa网络来说效率低下或过于重量级。FerryLink因此提出了一种结合RSSI采样和信道活动检测(CAD)的新型轮渡机制,以适应LoRa的多个正交传输参数。为了减少能量开销,FerryLink利用约定窗口在两个耦合节点之间进行粗粒度传输同步。最后,FerryLink利用上行和下行信号的正交性来避免由于轮渡机制造成的数据冗余,保持与原始lpwan相当的容量。我们用商用现成的硬件在lorawan之上构建FerryLink。广泛的评估结果表明,FerryLink与原有的LoRaWAN相比,有效地提高了LoRa节点的分组分发率(PDR)(95%以上),减少了2倍的能量开销,通信范围增加了50%。
{"title":"FerryLink: Combating Link Degradation for Practical LPWAN Deployments","authors":"Jing Yang, Zhenqiang Xu, Jiliang Wang","doi":"10.1109/ICPADS53394.2021.00077","DOIUrl":"https://doi.org/10.1109/ICPADS53394.2021.00077","url":null,"abstract":"Low-Power Wide-Area Networks (LPWANs) have been shown as a promising technique to provide long-range low-power communication for large-scale IoT devices. In this paper, however, we show the poor performance of LoRa network due to its link diversity in macro- and micro- scope through one-month measurements in an area of $2.2 kmtimes 1.5 km$. We present FerryLink, which exploits such link diversity and leverages peer nodes to ferry data of weak links, to combat performance degradation. Traditional arts (e.g., building multi-hop networks) are inefficient or too heavyweight for the current star-topology-based LoRa network. FerryLink thus proposes a novel ferry mechanism combining RSSI sampling and Channel Activity detection(CAD) to suit multiple orthogonal transmission parameters of LoRa. To reduce energy overhead, FerryLink leverages convention windows for coarse-grained transmission synchronization between two coupled nodes. Finally, FerryLink utilizes the orthogonality of uplink and downlink signals to avoid data redundancy due to the ferry mechanism, maintaining comparable capacity with original LPWANs. We build FerryLink on top of LoRaWANwith commercial off-the-shelf hardware. The extensive evaluation results show that FerryLink effectively improves the packet delivery rate (PDR) of LoRa nodes (to over 95%), achieves 2x less energy overhead, and increases communication range by 50% compared with the original LoRaWAN.","PeriodicalId":309508,"journal":{"name":"2021 IEEE 27th International Conference on Parallel and Distributed Systems (ICPADS)","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129795986","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Dynamic Push for HTTP Adaptive Streaming with Deep Reinforcement Learning 基于深度强化学习的HTTP自适应流动态推送
Pub Date : 2021-12-01 DOI: 10.1109/ICPADS53394.2021.00112
Haipeng Du, Danfu Yuan, Weizhan Zhang, Q. Zheng
HTTP adaptive streaming (HAS) has revolutionized video distribution over the Internet due to its prominent benefit of outstanding quality of experience (QoE). Due to the pull-based nature of HTTP/1.1, the client must make requests for each segment. This usually causes high request overhead and low bandwidth utilization and finally reduces QoE. Currently, research into the HAS adaptive bitrate algorithm typically focuses on the server-push feature introduced in the new HTTP standard, which enables the client to receive multiple segments with a single request. Every time a request is sent, the client must simultaneously make decisions on the number of segments the server should push and the bitrate of these future segments. As the decision space complexity increases, existing rule-based strategies inevitably fail to achieve optimal performance. In this paper, we present D-Push, an HAS framework that combines deep reinforcement learning (DRL) techniques. Instead of relying on inaccurate assumptions about the environment and network capacity variation models, D-Push trains a DRL model and makes decisions by exploiting the QoE of past decisions through the training process and adapts to a wide range of highly dynamic environments. The experimental results show that D-Push outperforms the existing state-of-the-art algorithm by 12%-24% in terms of the average QoE.
HTTP自适应流(HAS)由于其卓越的体验质量(QoE)的显著优势,已经彻底改变了互联网上的视频分发。由于HTTP/1.1基于拉的特性,客户端必须为每个段发出请求。这通常会导致高请求开销和低带宽利用率,最终降低QoE。目前,对HAS自适应比特率算法的研究主要集中在新HTTP标准中引入的服务器推送功能上,该功能使客户端能够通过单个请求接收多个段。每次发送请求时,客户端必须同时决定服务器应该推送的段的数量和这些未来段的比特率。随着决策空间复杂性的增加,现有的基于规则的策略不可避免地无法达到最优性能。在本文中,我们提出了D-Push,一个结合了深度强化学习(DRL)技术的HAS框架。D-Push不依赖于对环境和网络容量变化模型的不准确假设,而是训练一个DRL模型,并通过训练过程利用过去决策的QoE来做出决策,并适应大范围的高动态环境。实验结果表明,D-Push在平均QoE方面比现有最先进的算法高出12%-24%。
{"title":"Dynamic Push for HTTP Adaptive Streaming with Deep Reinforcement Learning","authors":"Haipeng Du, Danfu Yuan, Weizhan Zhang, Q. Zheng","doi":"10.1109/ICPADS53394.2021.00112","DOIUrl":"https://doi.org/10.1109/ICPADS53394.2021.00112","url":null,"abstract":"HTTP adaptive streaming (HAS) has revolutionized video distribution over the Internet due to its prominent benefit of outstanding quality of experience (QoE). Due to the pull-based nature of HTTP/1.1, the client must make requests for each segment. This usually causes high request overhead and low bandwidth utilization and finally reduces QoE. Currently, research into the HAS adaptive bitrate algorithm typically focuses on the server-push feature introduced in the new HTTP standard, which enables the client to receive multiple segments with a single request. Every time a request is sent, the client must simultaneously make decisions on the number of segments the server should push and the bitrate of these future segments. As the decision space complexity increases, existing rule-based strategies inevitably fail to achieve optimal performance. In this paper, we present D-Push, an HAS framework that combines deep reinforcement learning (DRL) techniques. Instead of relying on inaccurate assumptions about the environment and network capacity variation models, D-Push trains a DRL model and makes decisions by exploiting the QoE of past decisions through the training process and adapts to a wide range of highly dynamic environments. The experimental results show that D-Push outperforms the existing state-of-the-art algorithm by 12%-24% in terms of the average QoE.","PeriodicalId":309508,"journal":{"name":"2021 IEEE 27th International Conference on Parallel and Distributed Systems (ICPADS)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130930378","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Collaborative Framework of Cloud Transcoding and Distribution Supporting Cost-Efficient Crowdsourced Live Streaming 支持高性价比众包直播的云转码和分发协作框架
Pub Date : 2021-12-01 DOI: 10.1109/ICPADS53394.2021.00122
Jiannan Zheng, Haitao Zhang, Yilin Jin, Huadong Ma
With the rapid development of high-speed Internet access and popularization of high-performance smart devices, past decade has witnessed the great development of crowdsourced live streaming (CLS) service. Transcoding and video distribution are essential in CLS service to guarantee viewer engagement. Large CLS systems gradually migrate their services to multi-cloud platforms. However, highly dynamic viewers' requests influence transcoding and CDN distribution decisions, eventually lead to fluctuation in QoE and increase in operational cost. It is challenging for the CLS system to serve viewer's requests in multi-cloud platforms with fluctuation in cloud transcoding and distribution performance. In this paper, we propose a collaborative framework of cloud transcoding and distribution supporting CLS service. First, we define cost model and QoE model in multi-cloud platforms, comprehensively considering cloud transcoding and distribution. Second, we propose a collaborative cost-efficient approach based on multi-agent decision model. We use a G-Greedy exploration approach to learn what actions to take by exploration and exploitation based on the state of current environment. The trace-driven experiments demonstrate that our proposed approach is cost-efficient and QoE-maintained and can reduce operational cost compared with alternatives (5.37%-21.21%) while maintaining QoE of viewers.
随着高速互联网接入的快速发展和高性能智能设备的普及,过去十年见证了众包直播(CLS)服务的巨大发展。转换编码和视频分发在CLS服务中是必不可少的,以保证观众的参与。大型CLS系统逐渐将其服务迁移到多云平台。然而,高度动态的观众请求会影响转码和CDN分发决策,最终导致QoE的波动和运营成本的增加。由于云转码和分发性能的波动,CLS系统在多云平台上满足观看者的请求是一个挑战。在本文中,我们提出了一个支持CLS服务的云转码和分发协作框架。首先,我们定义了多云平台下的成本模型和QoE模型,综合考虑了云转码和云分布。其次,我们提出了一种基于多智能体决策模型的协同成本效益方法。我们使用G-Greedy探索方法,根据当前环境的状态来学习通过探索和开发采取什么行动。跟踪驱动实验表明,我们提出的方法具有成本效益和QoE保持,与其他方法相比,在保持观众QoE的情况下,可以降低运营成本(5.37%-21.21%)。
{"title":"Collaborative Framework of Cloud Transcoding and Distribution Supporting Cost-Efficient Crowdsourced Live Streaming","authors":"Jiannan Zheng, Haitao Zhang, Yilin Jin, Huadong Ma","doi":"10.1109/ICPADS53394.2021.00122","DOIUrl":"https://doi.org/10.1109/ICPADS53394.2021.00122","url":null,"abstract":"With the rapid development of high-speed Internet access and popularization of high-performance smart devices, past decade has witnessed the great development of crowdsourced live streaming (CLS) service. Transcoding and video distribution are essential in CLS service to guarantee viewer engagement. Large CLS systems gradually migrate their services to multi-cloud platforms. However, highly dynamic viewers' requests influence transcoding and CDN distribution decisions, eventually lead to fluctuation in QoE and increase in operational cost. It is challenging for the CLS system to serve viewer's requests in multi-cloud platforms with fluctuation in cloud transcoding and distribution performance. In this paper, we propose a collaborative framework of cloud transcoding and distribution supporting CLS service. First, we define cost model and QoE model in multi-cloud platforms, comprehensively considering cloud transcoding and distribution. Second, we propose a collaborative cost-efficient approach based on multi-agent decision model. We use a G-Greedy exploration approach to learn what actions to take by exploration and exploitation based on the state of current environment. The trace-driven experiments demonstrate that our proposed approach is cost-efficient and QoE-maintained and can reduce operational cost compared with alternatives (5.37%-21.21%) while maintaining QoE of viewers.","PeriodicalId":309508,"journal":{"name":"2021 IEEE 27th International Conference on Parallel and Distributed Systems (ICPADS)","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131502174","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2021 IEEE 27th International Conference on Parallel and Distributed Systems (ICPADS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1