首页 > 最新文献

[1992] Proceedings of the Fourth IEEE Symposium on Parallel and Distributed Processing最新文献

英文 中文
Multiple tree quorum algorithm for replica control in distributed database systems 分布式数据库系统中副本控制的多树仲裁算法
S. M. Chung, Cailin Cao
A novel replica control algorithm, called the multiple tree quorum (MTQ) algorithm, is proposed to manage replicated data in distributed database systems. This algorithm provides a high availability for read and write operations by imposing a logical structure of multiple trees on data copies. With the MTQ algorithm, a read operation is limited to a couple of data copies, and a write operation is allowed as long as the majority of the roots of the trees and the majority of the children of each node selected are available. Compared to other algorithms, the MTQ requires lower message cost for an operation while providing higher availability.<>
针对分布式数据库系统中的复制数据管理问题,提出了一种新的副本控制算法——多树仲裁算法(MTQ)。该算法通过在数据副本上施加多树的逻辑结构,为读写操作提供了高可用性。使用MTQ算法,读取操作仅限于几个数据副本,只要树的大多数根和所选每个节点的大多数子节点可用,就允许写操作。与其他算法相比,MTQ需要更低的操作消息成本,同时提供更高的可用性。
{"title":"Multiple tree quorum algorithm for replica control in distributed database systems","authors":"S. M. Chung, Cailin Cao","doi":"10.1109/SPDP.1992.242733","DOIUrl":"https://doi.org/10.1109/SPDP.1992.242733","url":null,"abstract":"A novel replica control algorithm, called the multiple tree quorum (MTQ) algorithm, is proposed to manage replicated data in distributed database systems. This algorithm provides a high availability for read and write operations by imposing a logical structure of multiple trees on data copies. With the MTQ algorithm, a read operation is limited to a couple of data copies, and a write operation is allowed as long as the majority of the roots of the trees and the majority of the children of each node selected are available. Compared to other algorithms, the MTQ requires lower message cost for an operation while providing higher availability.<<ETX>>","PeriodicalId":265469,"journal":{"name":"[1992] Proceedings of the Fourth IEEE Symposium on Parallel and Distributed Processing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1992-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128538419","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A balanced layered allocation scheme for hypercube based dataflow systems 基于超立方体数据流系统的均衡分层分配方案
Vincent R. Freytag, Ben Lee, A. Hurson
The authors propose a method called the balanced layered allocation scheme (BLAS) which utilizes heuristic rules, to find a balance between computation and communication costs in hypercube based dataflow systems. The central idea of this method is to arrange the nodes of a dataflow graph into layers that have a one-to-one correspondence with processors in a hypercube. During the allocation, CP (critical path) and LDP (longest directed path) heuristics determine the set of nodes which are to be assigned to processors. Each set of nodes is assigned in an iterative fashion to every possible layer. In this manner the effects of execution times and communication costs can be weighted against each other for every possible layer assignment. Sets of nodes are then assigned to the layer that yields the earliest completion time of the program. Simulation studies indicate that the proposed allocation scheme is effective in reducing communication overhead and thus the overall execution time of a program distribution on a hypercube dataflow computer. Overall, the BLAS showed promising improvements over the VL (vertically layered) allocation scheme.<>
作者提出了一种基于启发式规则的均衡分层分配方案(BLAS),以在基于超立方体的数据流系统中找到计算和通信成本之间的平衡。该方法的中心思想是将数据流图的节点排列到与超立方体中的处理器具有一对一对应关系的层中。在分配过程中,CP(关键路径)和LDP(最长有向路径)启发式方法确定要分配给处理器的节点集。每个节点集以迭代的方式分配给每个可能的层。通过这种方式,对于每个可能的层分配,执行时间和通信成本的影响可以相互加权。然后将节点集分配给产生程序最早完成时间的层。仿真研究表明,所提出的分配方案可以有效地减少通信开销,从而减少程序分布在超立方体数据流计算机上的总体执行时间。总体而言,与垂直分层分配方案相比,BLAS显示出有希望的改进。
{"title":"A balanced layered allocation scheme for hypercube based dataflow systems","authors":"Vincent R. Freytag, Ben Lee, A. Hurson","doi":"10.1109/SPDP.1992.242725","DOIUrl":"https://doi.org/10.1109/SPDP.1992.242725","url":null,"abstract":"The authors propose a method called the balanced layered allocation scheme (BLAS) which utilizes heuristic rules, to find a balance between computation and communication costs in hypercube based dataflow systems. The central idea of this method is to arrange the nodes of a dataflow graph into layers that have a one-to-one correspondence with processors in a hypercube. During the allocation, CP (critical path) and LDP (longest directed path) heuristics determine the set of nodes which are to be assigned to processors. Each set of nodes is assigned in an iterative fashion to every possible layer. In this manner the effects of execution times and communication costs can be weighted against each other for every possible layer assignment. Sets of nodes are then assigned to the layer that yields the earliest completion time of the program. Simulation studies indicate that the proposed allocation scheme is effective in reducing communication overhead and thus the overall execution time of a program distribution on a hypercube dataflow computer. Overall, the BLAS showed promising improvements over the VL (vertically layered) allocation scheme.<<ETX>>","PeriodicalId":265469,"journal":{"name":"[1992] Proceedings of the Fourth IEEE Symposium on Parallel and Distributed Processing","volume":"83 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1992-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123043753","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Modeling resource contention among distributed periodic processes 分布式周期性进程之间的资源争用建模
L. Welch, A. Stoyen, T. Marlowe
A framework for analyzing resource usage of distributed periodic processes specified in RT-Chart is provided. Functions for reasoning about the quality of an assignment of processes to computation nodes are developed, and a novel model of resource use costs and communication costs is shown. The model allows reasoning about the rates of progress each action makes when using its resources, thus modeling contention among periodic processes sharing a resource. The computed rates of progress are accurate since impossible sources of resource contention are removed. These are useful techniques for assignment algorithms (such as simulated annealing) that repeatedly reassign actions until a close-to-optimal solution is obtained.<>
给出了一种分析分布式周期过程资源使用情况的框架。开发了过程分配给计算节点的质量推理函数,并给出了一种新的资源使用成本和通信成本模型。该模型允许对每个操作在使用其资源时的进度进行推理,从而对共享资源的周期性进程之间的争用进行建模。计算的进度率是准确的,因为不可能的资源争用源被删除了。这些都是分配算法(如模拟退火)的有用技术,可以反复重新分配操作,直到获得接近最优的解决方案
{"title":"Modeling resource contention among distributed periodic processes","authors":"L. Welch, A. Stoyen, T. Marlowe","doi":"10.1109/SPDP.1992.242716","DOIUrl":"https://doi.org/10.1109/SPDP.1992.242716","url":null,"abstract":"A framework for analyzing resource usage of distributed periodic processes specified in RT-Chart is provided. Functions for reasoning about the quality of an assignment of processes to computation nodes are developed, and a novel model of resource use costs and communication costs is shown. The model allows reasoning about the rates of progress each action makes when using its resources, thus modeling contention among periodic processes sharing a resource. The computed rates of progress are accurate since impossible sources of resource contention are removed. These are useful techniques for assignment algorithms (such as simulated annealing) that repeatedly reassign actions until a close-to-optimal solution is obtained.<<ETX>>","PeriodicalId":265469,"journal":{"name":"[1992] Proceedings of the Fourth IEEE Symposium on Parallel and Distributed Processing","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1992-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115220460","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Event scheduling in window based parallel simulation schemes 基于窗口的并行仿真方案中的事件调度
R. Ayani, H. Rajaei
The authors study the impact of load balancing on the performance of a window-based parallel simulation scheme. They show the impact of four scheduling policies: static, dynamic, majority, and longest window. For simple simulation problems where there are a few events and the events need fairly short computation, i.e., small event granularity, the first three policies perform fairly close to each other. However, for large simulation problems where millions of events exist and the event granularity is large, majority scheduling is a better choice. For instance, if each event requires 18.3 ms processing time, majority scheduling performs 38.5% better than dynamic scheduling and 43.6% better than static scheduling.<>
研究了负载均衡对一种基于窗口的并行仿真方案性能的影响。它们显示了四种调度策略的影响:静态、动态、多数和最长窗口。对于简单的模拟问题,其中有几个事件,事件需要相当短的计算,即,小的事件粒度,前三个策略执行相当接近彼此。然而,对于存在数百万个事件且事件粒度较大的大型模拟问题,多数调度是更好的选择。例如,如果每个事件需要18.3 ms的处理时间,多数调度比动态调度好38.5%,比静态调度好43.6%。
{"title":"Event scheduling in window based parallel simulation schemes","authors":"R. Ayani, H. Rajaei","doi":"10.1109/SPDP.1992.242763","DOIUrl":"https://doi.org/10.1109/SPDP.1992.242763","url":null,"abstract":"The authors study the impact of load balancing on the performance of a window-based parallel simulation scheme. They show the impact of four scheduling policies: static, dynamic, majority, and longest window. For simple simulation problems where there are a few events and the events need fairly short computation, i.e., small event granularity, the first three policies perform fairly close to each other. However, for large simulation problems where millions of events exist and the event granularity is large, majority scheduling is a better choice. For instance, if each event requires 18.3 ms processing time, majority scheduling performs 38.5% better than dynamic scheduling and 43.6% better than static scheduling.<<ETX>>","PeriodicalId":265469,"journal":{"name":"[1992] Proceedings of the Fourth IEEE Symposium on Parallel and Distributed Processing","volume":"15 10","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1992-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132118815","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Tolerating faults in a mesh with a row of spare nodes 在有一排备用节点的网格中容忍故障
Jehoshua Bruck, R. Cypher, C. T. Ho
The authors present an efficient method for tolerating faults in a two-dimensional mesh architecture. The approach is based on adding spare components (nodes) and extra links (edges) such that the resulting architecture can be reconfigured as a mesh in the presence of faults. The cost of the fault-tolerant mesh architecture is optimized by adding about one row of redundant nodes in addition to a set of k spare nodes (while tolerating up to k node faults) and minimizing the number of links per node. The results are surprisingly efficient and seem to be practical for small values of k. The degree of the fault-tolerant architecture is k+5 for odd k, and k+6 for even k. The results can be generalized to d-dimensional meshes such that the number of spare nodes is less than the length of the shortest axis plus k, and the degree of the fault-tolerant mesh is (d-1)k+d+3 when k is odd and (d-1)k+2d+2 when k is even.<>
提出了一种有效的二维网格结构容错方法。该方法基于添加备用组件(节点)和额外链接(边),以便在存在故障的情况下将生成的体系结构重新配置为网格。通过在一组k个备用节点的基础上增加约一行冗余节点(同时允许最多k个节点故障)并最小化每个节点的链路数量,优化了容错网格体系结构的成本。结果令人惊讶地有效,并且对于较小的k值似乎是实用的。对于奇数k,容错架构的程度为k+5,对于偶数k,容错架构的程度为k+6。结果可以推广到d维网格,使得备用节点的数量小于最短轴的长度加上k,当k是奇数时,容错网格的程度为(d-1)k+d+3,当k是偶数时,容错网格的程度为(d-1)k+2d+2。
{"title":"Tolerating faults in a mesh with a row of spare nodes","authors":"Jehoshua Bruck, R. Cypher, C. T. Ho","doi":"10.1109/SPDP.1992.242768","DOIUrl":"https://doi.org/10.1109/SPDP.1992.242768","url":null,"abstract":"The authors present an efficient method for tolerating faults in a two-dimensional mesh architecture. The approach is based on adding spare components (nodes) and extra links (edges) such that the resulting architecture can be reconfigured as a mesh in the presence of faults. The cost of the fault-tolerant mesh architecture is optimized by adding about one row of redundant nodes in addition to a set of k spare nodes (while tolerating up to k node faults) and minimizing the number of links per node. The results are surprisingly efficient and seem to be practical for small values of k. The degree of the fault-tolerant architecture is k+5 for odd k, and k+6 for even k. The results can be generalized to d-dimensional meshes such that the number of spare nodes is less than the length of the shortest axis plus k, and the degree of the fault-tolerant mesh is (d-1)k+d+3 when k is odd and (d-1)k+2d+2 when k is even.<<ETX>>","PeriodicalId":265469,"journal":{"name":"[1992] Proceedings of the Fourth IEEE Symposium on Parallel and Distributed Processing","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1992-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124361695","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
The Scalable Tree Protocol-a cache coherence approach for large-scale multiprocessors 可扩展树协议-大规模多处理器的缓存一致性方法
H. Nilsson, P. Stenström
The problem of cache coherence in large-scale shared-memory multiprocessors has been addressed using directory-schemes. Two problems arise when the number of processors increases: the network latency increases and the implementation cost must be kept acceptable. The authors present a tree-based cache coherence protocol called the scalable tree protocol (STP). They show that it can be implemented at a reasonable implementation cost and that the write latency is logarithmic to the size of the sharing set. How to maintain an optimal tree structure and how to handle replacements efficiently are critical issues the authors address for this type of protocol. They compare the performance of the STP with that of the scalable coherent interface (SCI) (IEEE standard P1596) by considering a classical matrix-oriented algorithm targeted for large-scale parallel processing. They show that the STP manages to reduce the execution time considerably by reducing the write latency.<>
利用目录模式解决了大规模共享内存多处理器中的缓存一致性问题。当处理器数量增加时,会出现两个问题:网络延迟增加,实现成本必须保持在可接受的范围内。作者提出了一种基于树的缓存一致性协议,称为可扩展树协议(STP)。他们表明,它可以以合理的实现成本实现,并且写入延迟与共享集的大小成对数关系。如何保持一个最优的树结构和如何有效地处理替换是作者为这种类型的协议解决的关键问题。他们通过考虑针对大规模并行处理的经典面向矩阵算法,将STP的性能与可扩展相干接口(SCI) (IEEE标准P1596)的性能进行了比较。他们表明STP通过减少写延迟来大大减少执行时间。
{"title":"The Scalable Tree Protocol-a cache coherence approach for large-scale multiprocessors","authors":"H. Nilsson, P. Stenström","doi":"10.1109/SPDP.1992.242703","DOIUrl":"https://doi.org/10.1109/SPDP.1992.242703","url":null,"abstract":"The problem of cache coherence in large-scale shared-memory multiprocessors has been addressed using directory-schemes. Two problems arise when the number of processors increases: the network latency increases and the implementation cost must be kept acceptable. The authors present a tree-based cache coherence protocol called the scalable tree protocol (STP). They show that it can be implemented at a reasonable implementation cost and that the write latency is logarithmic to the size of the sharing set. How to maintain an optimal tree structure and how to handle replacements efficiently are critical issues the authors address for this type of protocol. They compare the performance of the STP with that of the scalable coherent interface (SCI) (IEEE standard P1596) by considering a classical matrix-oriented algorithm targeted for large-scale parallel processing. They show that the STP manages to reduce the execution time considerably by reducing the write latency.<<ETX>>","PeriodicalId":265469,"journal":{"name":"[1992] Proceedings of the Fourth IEEE Symposium on Parallel and Distributed Processing","volume":"123 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1992-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134623294","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 44
Time sharing in hypercube multiprocessors 超立方体多处理器中的时间共享
E. Filho, V. Barbosa
In hypercube multiprocessors, multiple users are normally supported by dividing the cube into subcubes of different dimensions. A user request for a subcube may be denied depending on how other subcubes were previously allocated or because the allocation algorithm fails to recognize an existing free subcube that would satisfy the request. In both cases, the main consequence is a reduction in system utilization. To support multiple users while avoiding such problems, the authors propose to multiplex all the processors in the cube among the users. In this way, one can get full system utilization and offer all the resources of the system to the users. The authors have conducted experiments with this new approach on an Intel iPSC/860 system, comparing its performance with the one obtained in the conventional cube-partitioning approach. The results show that, for computationally intensive applications, the average execution time per user when multiplexing the processors is generally comparable to the same average when allocating subcubes to the users, and often significantly lower.<>
在超多维数据集多处理器中,通常通过将多维数据集划分为不同维度的子多维数据集来支持多个用户。用户对子多维数据集的请求可能会被拒绝,这取决于以前如何分配其他子多维数据集,或者因为分配算法无法识别满足请求的现有空闲子多维数据集。在这两种情况下,主要后果都是系统利用率的降低。为了在支持多个用户的同时避免此类问题,作者建议将多维数据集中的所有处理器多路分配给用户。这样可以充分利用系统,将系统的所有资源提供给用户。作者在Intel iPSC/860系统上对这种新方法进行了实验,并将其性能与传统的立方体分区方法进行了比较。结果表明,对于计算密集型应用程序,多路复用处理器时每个用户的平均执行时间通常与为用户分配子数据集时相同的平均执行时间相当,并且通常要低得多。
{"title":"Time sharing in hypercube multiprocessors","authors":"E. Filho, V. Barbosa","doi":"10.1109/SPDP.1992.242724","DOIUrl":"https://doi.org/10.1109/SPDP.1992.242724","url":null,"abstract":"In hypercube multiprocessors, multiple users are normally supported by dividing the cube into subcubes of different dimensions. A user request for a subcube may be denied depending on how other subcubes were previously allocated or because the allocation algorithm fails to recognize an existing free subcube that would satisfy the request. In both cases, the main consequence is a reduction in system utilization. To support multiple users while avoiding such problems, the authors propose to multiplex all the processors in the cube among the users. In this way, one can get full system utilization and offer all the resources of the system to the users. The authors have conducted experiments with this new approach on an Intel iPSC/860 system, comparing its performance with the one obtained in the conventional cube-partitioning approach. The results show that, for computationally intensive applications, the average execution time per user when multiplexing the processors is generally comparable to the same average when allocating subcubes to the users, and often significantly lower.<<ETX>>","PeriodicalId":265469,"journal":{"name":"[1992] Proceedings of the Fourth IEEE Symposium on Parallel and Distributed Processing","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1992-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122357629","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
A fast sort using parallelism within memory 在内存中使用并行性的快速排序
C. Leopold
The author models the internal structure of memory by a tree, where nodes represent memory modules (like cache, disks), and edges represent buses between them. The modules have smaller access time, capacity, and block size the nearer they are to the root. All buses may transmit blocks of data in parallel. The author gives a deterministic sorting algorithm based on greed-sort. Its running time is shown to be optimal up to a constant factor. The bound implies the number of parallel modules necessary at each hierarchy level to overcome the I/O bottlenecks of sorting. The proposed algorithm also applies to the less general models UMH (uniform memory hierarchies) and P-UMH.<>
作者通过树来建模内存的内部结构,其中节点表示内存模块(如缓存、磁盘),边表示它们之间的总线。模块越靠近根,其访问时间、容量和块大小就越小。所有总线都可以并行传输数据块。给出了一种基于贪婪排序的确定性排序算法。它的运行时间被证明是最优的,直到一个常数因子。该边界表示在每个层次结构级别上克服排序的I/O瓶颈所需的并行模块的数量。提出的算法也适用于不太通用的模型UMH(统一内存层次结构)和P-UMH。
{"title":"A fast sort using parallelism within memory","authors":"C. Leopold","doi":"10.1109/SPDP.1992.242727","DOIUrl":"https://doi.org/10.1109/SPDP.1992.242727","url":null,"abstract":"The author models the internal structure of memory by a tree, where nodes represent memory modules (like cache, disks), and edges represent buses between them. The modules have smaller access time, capacity, and block size the nearer they are to the root. All buses may transmit blocks of data in parallel. The author gives a deterministic sorting algorithm based on greed-sort. Its running time is shown to be optimal up to a constant factor. The bound implies the number of parallel modules necessary at each hierarchy level to overcome the I/O bottlenecks of sorting. The proposed algorithm also applies to the less general models UMH (uniform memory hierarchies) and P-UMH.<<ETX>>","PeriodicalId":265469,"journal":{"name":"[1992] Proceedings of the Fourth IEEE Symposium on Parallel and Distributed Processing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1992-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114925674","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Fault-tolerant embeddings of rings, meshes, and tori in hypercubes 超立方体中环、网格和环面的容错嵌入
Alexander Wang, R. Cypher
The authors study the ability of the hypercube to implement algorithms with ring, mesh, and torus communication patterns when the hypercube contains faults. The primary result is a fault-free embedding of the longest possible ring into an n-cube with at most (n-h(n)) even faulty nodes and (n-h(n)) odd faulty nodes, where h(n) is a function such that h(n)=O( square root n log n). Given the above bounds on the parities of the faults, the result obtained improved upon previous results both in the number of faults that are tolerated and in the length of the ring that is embedded. In addition, the result leads to improved bounds for fault-free embeddings of meshes and tori into faulty hypercubes.<>
作者研究了当超立方体包含故障时,超立方体实现环形、网格和环面通信模式算法的能力。主要结果是将最长可能的环无故障嵌入到最多(n-h(n))个偶数故障节点和(n-h(n))个奇数故障节点的n-立方体中,其中h(n)是h(n)=O(平方根n log n)的函数。给定上述故障对偶的界限,所获得的结果在可容忍的故障数量和嵌入的环长度方面都比先前的结果有改进。此外,该结果还改善了网格和环面无故障嵌入到故障超立方体中的边界。
{"title":"Fault-tolerant embeddings of rings, meshes, and tori in hypercubes","authors":"Alexander Wang, R. Cypher","doi":"10.1109/SPDP.1992.242767","DOIUrl":"https://doi.org/10.1109/SPDP.1992.242767","url":null,"abstract":"The authors study the ability of the hypercube to implement algorithms with ring, mesh, and torus communication patterns when the hypercube contains faults. The primary result is a fault-free embedding of the longest possible ring into an n-cube with at most (n-h(n)) even faulty nodes and (n-h(n)) odd faulty nodes, where h(n) is a function such that h(n)=O( square root n log n). Given the above bounds on the parities of the faults, the result obtained improved upon previous results both in the number of faults that are tolerated and in the length of the ring that is embedded. In addition, the result leads to improved bounds for fault-free embeddings of meshes and tori into faulty hypercubes.<<ETX>>","PeriodicalId":265469,"journal":{"name":"[1992] Proceedings of the Fourth IEEE Symposium on Parallel and Distributed Processing","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1992-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116592687","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Hierarchical shuffle-exchange and de Bruijn networks 等级洗牌交换和德布鲁因网络
R. Cypher, J. Sanz
The authors present two new classes of interconnection networks for SIMD (single-instruction multiple-data) computers, namely, the hierarchical shuffle-exchange (HSE) and hierarchical de Bruijn (HdB) networks. These networks are efficient in implementing a wide range of algorithms, including all of those in the classes ascend and descend. The networks are highly regular and scalable, and well-suited to VLSI implementation. In addition, they can be adjusted to match the pin limitations imposed by the packaging technology. The authors compare the HSE and HdB networks with hypercube, 2-D mesh, 3-D mesh, shuffle-exchange, hypernet, de Bruijn, and cube-connected cycles networks. The HSE and HdB networks are shown to have advantages in terms of regularity, scalability, and performance.<>
作者提出了两种新的SIMD(单指令多数据)计算机互连网络,即分层shuffle-exchange (HSE)和分层de Bruijn (HdB)网络。这些网络在实现广泛的算法方面是有效的,包括所有在上升和下降类中的算法。该网络具有高度的规则性和可扩展性,非常适合VLSI实现。此外,它们可以调整以匹配封装技术施加的引脚限制。作者将HSE和HdB网络与超立方体、二维网格、三维网格、shuffle-exchange、超网络、de Bruijn和立方体连接循环网络进行了比较。HSE和HdB网络在规律性、可扩展性和性能方面具有优势。
{"title":"Hierarchical shuffle-exchange and de Bruijn networks","authors":"R. Cypher, J. Sanz","doi":"10.1109/SPDP.1992.242704","DOIUrl":"https://doi.org/10.1109/SPDP.1992.242704","url":null,"abstract":"The authors present two new classes of interconnection networks for SIMD (single-instruction multiple-data) computers, namely, the hierarchical shuffle-exchange (HSE) and hierarchical de Bruijn (HdB) networks. These networks are efficient in implementing a wide range of algorithms, including all of those in the classes ascend and descend. The networks are highly regular and scalable, and well-suited to VLSI implementation. In addition, they can be adjusted to match the pin limitations imposed by the packaging technology. The authors compare the HSE and HdB networks with hypercube, 2-D mesh, 3-D mesh, shuffle-exchange, hypernet, de Bruijn, and cube-connected cycles networks. The HSE and HdB networks are shown to have advantages in terms of regularity, scalability, and performance.<<ETX>>","PeriodicalId":265469,"journal":{"name":"[1992] Proceedings of the Fourth IEEE Symposium on Parallel and Distributed Processing","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1992-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125126244","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
[1992] Proceedings of the Fourth IEEE Symposium on Parallel and Distributed Processing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1