首页 > 最新文献

Proceedings of the Fifth Distributed Memory Computing Conference, 1990.最新文献

英文 中文
Evaluation of Dual Ported Memories from the Task Level 从任务层面评价双端口记忆
Pub Date : 1990-04-08 DOI: 10.1109/DMCC.1990.556267
Rutger F. H. Hofman
An architecture, which is a hybrid of local memory and shared memory, is described in this report: it uses dual ported memories (DPMs), each accessed by two processors. Each processor is connected to a number of DPMs. The profit that is gained by using a DPM as a shared memory between two processors appears from task allocation results: task transport costs are avoided when a task, newly created in DPM d by one of d’s two processors, is allocated to the other processor at d. For a number of task allocation strategies, simulation studies show that the fraction of the tasks that benefit from this optimisation decreases with the number of processors in the multiprocessor. For larger numbers of processors, this fraction is considerably higher than the fraction under random allocation.
本报告描述了一种混合了本地内存和共享内存的体系结构:它使用双端口内存(dpm),每个dpm由两个处理器访问。每个处理器都连接到多个dpm。使用DPM作为两个处理器之间的共享内存所获得的利润出现在任务分配结果中:当DPM d中新创建的任务由d的两个处理器之一分配给d的另一个处理器时,可以避免任务传输成本。对于许多任务分配策略,仿真研究表明,受益于这种优化的任务比例随着多处理器中的处理器数量的减少而减少。对于更大数量的处理器,这个分数远远高于随机分配下的分数。
{"title":"Evaluation of Dual Ported Memories from the Task Level","authors":"Rutger F. H. Hofman","doi":"10.1109/DMCC.1990.556267","DOIUrl":"https://doi.org/10.1109/DMCC.1990.556267","url":null,"abstract":"An architecture, which is a hybrid of local memory and shared memory, is described in this report: it uses dual ported memories (DPMs), each accessed by two processors. Each processor is connected to a number of DPMs. The profit that is gained by using a DPM as a shared memory between two processors appears from task allocation results: task transport costs are avoided when a task, newly created in DPM d by one of d’s two processors, is allocated to the other processor at d. For a number of task allocation strategies, simulation studies show that the fraction of the tasks that benefit from this optimisation decreases with the number of processors in the multiprocessor. For larger numbers of processors, this fraction is considerably higher than the fraction under random allocation.","PeriodicalId":204431,"journal":{"name":"Proceedings of the Fifth Distributed Memory Computing Conference, 1990.","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125575999","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
An Approach to Compiling Single-point Iterative Programs for Distributed Memory Computers 分布式存储计算机单点迭代程序的编译方法
Pub Date : 1990-04-08 DOI: 10.1109/DMCC.1990.556313
D. Socha
This paper proposes a scheme for compiling an important class of iterative algorithms into efficient code for distributed memory computers. The programmer provides a description of the problem in Spot: a data parallel SIMD language that uses iterations as the unit of synchronization and is based on grids of data points. The data parallel description is in terms of a single point of the data space, with implicit communication semantics, and a set of numerical boundary conditions. The compiler eliminates the need for multi-tasking by ‘(expanding” the single-point code into multiplepoint code that executes over rectangular regions of points. Using rectangle intersection and difference operations on these regions allows the compiler to automatically insert the required communication calls and to hide communication latency by overlapping comput at ion and communication. The multiple-point code may be specialized, at compile-time, to the size and shape of different allocations, or it may use table-driven for-loops to adapt, at run-time, to the shape and size of the allocations. We show how to generalize this strategy to produce code for the near-rectangular shaped allocations required for balanced partitionings of rectangular arrays.
本文提出了一种将一类重要的迭代算法编译成分布式存储计算机的高效代码的方案。程序员在Spot中提供了对问题的描述:一种数据并行SIMD语言,它使用迭代作为同步单元,并基于数据点网格。数据并行描述是根据数据空间的单点、隐式通信语义和一组数值边界条件进行的。编译器通过将单点代码“扩展”为在矩形点区域上执行的多点代码,消除了对多任务的需要。在这些区域上使用矩形相交和差分操作,可以使编译器自动插入所需的通信调用,并通过重叠计算和通信来隐藏通信延迟。多点代码可以在编译时专门用于不同分配的大小和形状,也可以使用表驱动的for循环在运行时适应分配的形状和大小。我们将展示如何推广此策略,以生成用于矩形数组均衡分区所需的近矩形分配的代码。
{"title":"An Approach to Compiling Single-point Iterative Programs for Distributed Memory Computers","authors":"D. Socha","doi":"10.1109/DMCC.1990.556313","DOIUrl":"https://doi.org/10.1109/DMCC.1990.556313","url":null,"abstract":"This paper proposes a scheme for compiling an important class of iterative algorithms into efficient code for distributed memory computers. The programmer provides a description of the problem in Spot: a data parallel SIMD language that uses iterations as the unit of synchronization and is based on grids of data points. The data parallel description is in terms of a single point of the data space, with implicit communication semantics, and a set of numerical boundary conditions. The compiler eliminates the need for multi-tasking by ‘(expanding” the single-point code into multiplepoint code that executes over rectangular regions of points. Using rectangle intersection and difference operations on these regions allows the compiler to automatically insert the required communication calls and to hide communication latency by overlapping comput at ion and communication. The multiple-point code may be specialized, at compile-time, to the size and shape of different allocations, or it may use table-driven for-loops to adapt, at run-time, to the shape and size of the allocations. We show how to generalize this strategy to produce code for the near-rectangular shaped allocations required for balanced partitionings of rectangular arrays.","PeriodicalId":204431,"journal":{"name":"Proceedings of the Fifth Distributed Memory Computing Conference, 1990.","volume":"87 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126014590","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Mapping Data to Processors in Distributed Memory Computations 分布式内存计算中数据到处理器的映射
Pub Date : 1990-04-08 DOI: 10.1109/DMCC.1990.556295
M. Rosing, R. P. Weaver
Abstract : The authors present a structured scheme for allowing a programmer to specify the mapping of data to distributed memory multiprocessors. This scheme lets the programmer specify information about communication patterns as well as information about distributing data structures onto processors (including partitioning with replication). This mapping scheme allows the user to map arrays of data to arrays of processors. The user specifies how each axis of the data structure is mapped onto an axis of the processor structure. This mapping may either be one to one or one to many depending on the parallelism, load balancing, and communication requirements. The authors discuss the basics of how this scheme is implemented in the DINO language, the areas in which it has worked well, the few areas in which there were significant problems, and some ideas for future improvements.
摘要:作者提出了一种结构化方案,允许程序员指定数据到分布式内存多处理器的映射。该方案允许程序员指定有关通信模式的信息以及有关将数据结构分布到处理器上的信息(包括带有复制的分区)。这种映射方案允许用户将数据数组映射到处理器数组。用户指定如何将数据结构的每个轴映射到处理器结构的轴上。这种映射可以是一对一的,也可以是一对多的,这取决于并行性、负载平衡和通信需求。作者讨论了该方案如何在DINO语言中实现的基础知识,它工作良好的领域,存在重大问题的少数领域,以及未来改进的一些想法。
{"title":"Mapping Data to Processors in Distributed Memory Computations","authors":"M. Rosing, R. P. Weaver","doi":"10.1109/DMCC.1990.556295","DOIUrl":"https://doi.org/10.1109/DMCC.1990.556295","url":null,"abstract":"Abstract : The authors present a structured scheme for allowing a programmer to specify the mapping of data to distributed memory multiprocessors. This scheme lets the programmer specify information about communication patterns as well as information about distributing data structures onto processors (including partitioning with replication). This mapping scheme allows the user to map arrays of data to arrays of processors. The user specifies how each axis of the data structure is mapped onto an axis of the processor structure. This mapping may either be one to one or one to many depending on the parallelism, load balancing, and communication requirements. The authors discuss the basics of how this scheme is implemented in the DINO language, the areas in which it has worked well, the few areas in which there were significant problems, and some ideas for future improvements.","PeriodicalId":204431,"journal":{"name":"Proceedings of the Fifth Distributed Memory Computing Conference, 1990.","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127401473","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
An Adaptive Multiscale Scheme for Real-Time Motion Field Estimation 一种实时运动场估计的自适应多尺度方案
Pub Date : 1990-04-08 DOI: 10.1109/DMCC.1990.555383
R. Battiti
The problem considered in this work is that of estimating the motion field (i.e. the projection of the velocity field onto the image plane) from a temporal sequence of images. Generic images contain different objects with diverse spatial frequencies and motion amplitudes. To deal with this complex environment in a fast and effective way, biological visual systems use parallel processing, visual channels at different resolutions and adaptive mechanisms. In this paper a new adaptive multiscale scheme is proposed, in which the spatial discretization scale is based on a local estimate of the errors involved. Considering the constraints for real-time operation, flexibility and portability, the scheme can be implemented on MIMD parallel computers with medium size grains with high efficiency. Tests with ray-traced and video-acquired images for different motion ranges show that this method produces a better estimation with respect to the homogeneous (no Gadap t ive) mult iscale met hod.
在这项工作中考虑的问题是从图像的时间序列中估计运动场(即速度场在图像平面上的投影)。通用图像包含不同空间频率和运动幅度的不同对象。为了快速有效地处理这种复杂的环境,生物视觉系统采用并行处理、不同分辨率的视觉通道和自适应机制。本文提出了一种新的自适应多尺度方案,该方案的空间离散尺度基于误差的局部估计。考虑到实时性、灵活性和可移植性的限制,该方案可以在中等粒度的MIMD并行计算机上高效实现。对不同运动范围的光线跟踪和视频采集图像进行的测试表明,该方法相对于均匀(无Gadap t - ive)多尺度方法产生了更好的估计。
{"title":"An Adaptive Multiscale Scheme for Real-Time Motion Field Estimation","authors":"R. Battiti","doi":"10.1109/DMCC.1990.555383","DOIUrl":"https://doi.org/10.1109/DMCC.1990.555383","url":null,"abstract":"The problem considered in this work is that of estimating the motion field (i.e. the projection of the velocity field onto the image plane) from a temporal sequence of images. Generic images contain different objects with diverse spatial frequencies and motion amplitudes. To deal with this complex environment in a fast and effective way, biological visual systems use parallel processing, visual channels at different resolutions and adaptive mechanisms. In this paper a new adaptive multiscale scheme is proposed, in which the spatial discretization scale is based on a local estimate of the errors involved. Considering the constraints for real-time operation, flexibility and portability, the scheme can be implemented on MIMD parallel computers with medium size grains with high efficiency. Tests with ray-traced and video-acquired images for different motion ranges show that this method produces a better estimation with respect to the homogeneous (no Gadap t ive) mult iscale met hod.","PeriodicalId":204431,"journal":{"name":"Proceedings of the Fifth Distributed Memory Computing Conference, 1990.","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126646154","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Distributed Memory Implementation of SISAL SISAL的分布式内存实现
Pub Date : 1990-04-08 DOI: 10.1109/DMCC.1990.556327
D. Grit
SISAL is a general-purpose applicative 11 anguage intended for use on both conventional aiid novel multiprocessor systems. In this paper we describe the port of a shared memory implemeni,ation to a distributed memory environment. A ni mber of issues are specifically addressed: the e~~aluation strategy, memory management, schedulinp , stream handling, and task synchronization.
SISAL是一种通用的应用语言,用于传统的和新型的多处理器系统。在本文中,我们描述了共享内存实现到分布式内存环境的移植。具体解决了许多问题:e~~求值策略、内存管理、调度、流处理和任务同步。
{"title":"A Distributed Memory Implementation of SISAL","authors":"D. Grit","doi":"10.1109/DMCC.1990.556327","DOIUrl":"https://doi.org/10.1109/DMCC.1990.556327","url":null,"abstract":"SISAL is a general-purpose applicative 11 anguage intended for use on both conventional aiid novel multiprocessor systems. In this paper we describe the port of a shared memory implemeni,ation to a distributed memory environment. A ni mber of issues are specifically addressed: the e~~aluation strategy, memory management, schedulinp , stream handling, and task synchronization.","PeriodicalId":204431,"journal":{"name":"Proceedings of the Fifth Distributed Memory Computing Conference, 1990.","volume":"73 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114182629","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Hot-Spot Performance of Single-Stage and Multistage Interconnection Networks 单级和多级互连网络的热点性能
Pub Date : 1990-04-08 DOI: 10.1109/DMCC.1990.556269
K. Gunter, E. Gehringer
{"title":"Hot-Spot Performance of Single-Stage and Multistage Interconnection Networks","authors":"K. Gunter, E. Gehringer","doi":"10.1109/DMCC.1990.556269","DOIUrl":"https://doi.org/10.1109/DMCC.1990.556269","url":null,"abstract":"","PeriodicalId":204431,"journal":{"name":"Proceedings of the Fifth Distributed Memory Computing Conference, 1990.","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126673213","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Reducing Inner Product Computation in the Parallel One-Sided Jacobi Algorithm 减少平行单侧Jacobi算法的内积计算
Pub Date : 1990-04-08 DOI: 10.1109/DMCC.1990.555398
C. Romine, K. Sigmon
{"title":"Reducing Inner Product Computation in the Parallel One-Sided Jacobi Algorithm","authors":"C. Romine, K. Sigmon","doi":"10.1109/DMCC.1990.555398","DOIUrl":"https://doi.org/10.1109/DMCC.1990.555398","url":null,"abstract":"","PeriodicalId":204431,"journal":{"name":"Proceedings of the Fifth Distributed Memory Computing Conference, 1990.","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"113977152","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient Serial and Parallel Subcube Recognition in Hypercubes 超立方体中高效的串行和并行子立方体识别
Pub Date : 1990-04-08 DOI: 10.1109/DMCC.1990.555363
S. Al-Bassam, H. El-Rewini, B. Bose, T. Lewis
We develop an efficient subcube recognition algorithm that recognizes all the possible subcubes. The algorithm is based on exploiting more subcubes at different levels of the buddy tree. In exploiting the different levels, the algorithm checks any subcube at most once. Moreover, many unavailable subcubes are not considered as candidates and hence not checked for availability. This makes the algorithm fast in recognizing the subcubes. The number of recognized subcubes, for different subcube sizes, can be easily adjusted by restricting the search level down the buddy tree. The previous known algorithms become a special case of this general approach. When one level is searched, this algorithm perfoms as the original buddy system. When two levels are searched, it will recognized the Same subcubes as the ones in [4] with a faster speed. When all the levels are searched, a complete subcube recognition is obtained. In a multi-processing system, each processor can execute this algorithm on a different tree. Using a given number of processors in a multi-processing system, we give a method of constructing the trees that maximizes the overall number of recognized subcubes. Finally, we introduce an allocation method "best fit" that reduces hypercube fragmentation. Simulation results and performance comparisons between this method and the traditional "first fit" are presented.
我们开发了一种有效的子立方体识别算法,可以识别所有可能的子立方体。该算法基于在伙伴树的不同层次上挖掘更多的子数据集。在利用不同的层次时,该算法最多检查一次任何子立方体。此外,许多不可用的子数据集不被视为候选数据集,因此不检查可用性。这使得该算法能够快速识别子数据集。对于不同大小的子多维数据集,可以通过限制好友树的搜索级别来轻松调整可识别的子多维数据集的数量。以前已知的算法成为这种通用方法的特例。当搜索一个关卡时,该算法执行原始的伙伴系统。当搜索两个级别时,它将以更快的速度识别与[4]中相同的子数据集。当搜索所有级别时,获得完整的子立方体识别。在多处理系统中,每个处理器可以在不同的树上执行该算法。在多处理系统中使用给定数量的处理器,我们给出了一种构造树的方法,该树可以最大限度地增加可识别子数据集的总数。最后,我们引入了一种减少超立方体碎片的“最佳拟合”分配方法。给出了该方法与传统的“首次拟合”方法的仿真结果和性能比较。
{"title":"Efficient Serial and Parallel Subcube Recognition in Hypercubes","authors":"S. Al-Bassam, H. El-Rewini, B. Bose, T. Lewis","doi":"10.1109/DMCC.1990.555363","DOIUrl":"https://doi.org/10.1109/DMCC.1990.555363","url":null,"abstract":"We develop an efficient subcube recognition algorithm that recognizes all the possible subcubes. The algorithm is based on exploiting more subcubes at different levels of the buddy tree. In exploiting the different levels, the algorithm checks any subcube at most once. Moreover, many unavailable subcubes are not considered as candidates and hence not checked for availability. This makes the algorithm fast in recognizing the subcubes. The number of recognized subcubes, for different subcube sizes, can be easily adjusted by restricting the search level down the buddy tree. The previous known algorithms become a special case of this general approach. When one level is searched, this algorithm perfoms as the original buddy system. When two levels are searched, it will recognized the Same subcubes as the ones in [4] with a faster speed. When all the levels are searched, a complete subcube recognition is obtained. In a multi-processing system, each processor can execute this algorithm on a different tree. Using a given number of processors in a multi-processing system, we give a method of constructing the trees that maximizes the overall number of recognized subcubes. Finally, we introduce an allocation method \"best fit\" that reduces hypercube fragmentation. Simulation results and performance comparisons between this method and the traditional \"first fit\" are presented.","PeriodicalId":204431,"journal":{"name":"Proceedings of the Fifth Distributed Memory Computing Conference, 1990.","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123955406","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Hypercubes for Critical Space Flight Command Operations 关键空间飞行指挥操作的超立方体
Pub Date : 1990-04-08 DOI: 10.1109/DMCC.1990.555355
J. Horvath, T. Tang, L. P. Perry, R. Cole, D.B. OIster, J. Zipse
Controlling interplanetary spacecraft and planning their activities, as currently practiced, requires massive amounts of computer time and personnel. To improve this situation, it is desired to use advanced computing to speed up and automate the commanding process. Several design and prototype efforts have been underway at JPL to understand the appropriate roles for concurrent processors in future interplanetary spacecraft operations. Here we report on an effort to identify likely candidates for parallelism among existing software systems that both generate commands to be sent to the spacecraft and simulate what the spacecraft will do with these commands when it receives them. We also describe promising results from efforts to create parallel prototypes of representative portions of these software systems on the JPL/Caltech Mark 111 hypercube.
控制行星际航天器并规划其活动,如目前所实行的,需要大量的计算机时间和人员。为了改善这种情况,需要使用先进的计算来加速和自动化指挥过程。喷气推进实验室正在进行一些设计和原型工作,以了解并发处理器在未来星际航天器操作中的适当作用。在这里,我们报告了一项努力,以确定现有软件系统中可能的并行候选,这些软件系统既生成要发送给航天器的命令,又模拟航天器在接收这些命令时将如何处理这些命令。我们还描述了在JPL/Caltech Mark 111超立方体上创建这些软件系统的代表性部分的并行原型的有希望的结果。
{"title":"Hypercubes for Critical Space Flight Command Operations","authors":"J. Horvath, T. Tang, L. P. Perry, R. Cole, D.B. OIster, J. Zipse","doi":"10.1109/DMCC.1990.555355","DOIUrl":"https://doi.org/10.1109/DMCC.1990.555355","url":null,"abstract":"Controlling interplanetary spacecraft and planning their activities, as currently practiced, requires massive amounts of computer time and personnel. To improve this situation, it is desired to use advanced computing to speed up and automate the commanding process. Several design and prototype efforts have been underway at JPL to understand the appropriate roles for concurrent processors in future interplanetary spacecraft operations. Here we report on an effort to identify likely candidates for parallelism among existing software systems that both generate commands to be sent to the spacecraft and simulate what the spacecraft will do with these commands when it receives them. We also describe promising results from efforts to create parallel prototypes of representative portions of these software systems on the JPL/Caltech Mark 111 hypercube.","PeriodicalId":204431,"journal":{"name":"Proceedings of the Fifth Distributed Memory Computing Conference, 1990.","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127664534","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Load Sharing In Hypercube Multicomputers In The Presence Of Node Failures 节点故障情况下超立方体多计算机的负载分担
Pub Date : 1990-04-08 DOI: 10.1109/DMCC.1990.556410
Yi-Chieh Chang, K. Shin
This paper discusses and analyzes two load sharing (LS) issues: adjusting preferred lists and implementing a fault-tolerant mechanism in the presence of node failures. In an early paper, we have proposed to order the nodes in each node's proximity into a preferred list for the purpose of load sharing in distributed real-time systems. The preferred list of each node is constructed in such a way that each node will be selected as the kth preferred node by one and only one other node. Such lists are proven to allow the tasks to be evenly distributed in a system. However, the presence of faulty nodes will destroy the original structure of a preferred list if the faulty nodes are simply skipped in the preferred list. An algorithm is therefore proposed to modify each preferred list to retain its original features regardless of the number of faulty nodes in the system. The communication overhead introduced by this algorithm is shown to be minimal. Based on the modified preferred lists, a simple fault-tolerant mechanism is implemented. Each node is equipped with a backup queue which 'The work reported in this paper was supported in part by the Office of Naval Research under contract N0001485-K-0122, and the NSF under grant DMC-8721492. Any opinions, findings, and recommendations expressed in this publication are those of the authors and do not necessarily reflect the view of the funding agencies. stores and updates the arriving/completing tasks at its most preferrecd node. Whenever a node becomes faulty, its m.ost preferred node will treat the tasks in the baxkup queue as externally axriving tasks. Our simulation results show that this approach, despite of the simplicity, can reduce the number of task losses dramatically, as compared to the approaches without any faulttolerant mechanism.
本文讨论并分析了负载共享(LS)的两个问题:调整首选列表和实现节点故障时的容错机制。在早期的一篇论文中,我们提出了在分布式实时系统中,为了负载共享的目的,将每个节点邻近的节点排序到一个首选列表中。构建每个节点的首选列表的方式是,每个节点将被一个且仅一个其他节点选择为第k个首选节点。这样的列表被证明允许任务在系统中均匀分布。但是,如果在首选列表中跳过故障节点,则故障节点的存在将破坏首选列表的原始结构。因此,提出了一种算法,无论系统中故障节点的数量如何,都可以修改每个首选列表以保留其原始特征。该算法带来的通信开销是最小的。基于修改后的首选列表,实现了简单的容错机制。每个节点都配备了一个备份队列,本文报告的工作部分由海军研究办公室根据合同N0001485-K-0122和NSF根据拨款DMC-8721492提供支持。本出版物中表达的任何观点、发现和建议均为作者的观点,并不一定反映资助机构的观点。在其最优选的节点上存储和更新到达/完成的任务。当一个节点发生故障时,它的最优节点将把备份队列中的任务视为外部到达的任务。我们的仿真结果表明,尽管这种方法简单,但与没有容错机制的方法相比,可以显着减少任务损失的数量。
{"title":"Load Sharing In Hypercube Multicomputers In The Presence Of Node Failures","authors":"Yi-Chieh Chang, K. Shin","doi":"10.1109/DMCC.1990.556410","DOIUrl":"https://doi.org/10.1109/DMCC.1990.556410","url":null,"abstract":"This paper discusses and analyzes two load sharing (LS) issues: adjusting preferred lists and implementing a fault-tolerant mechanism in the presence of node failures. In an early paper, we have proposed to order the nodes in each node's proximity into a preferred list for the purpose of load sharing in distributed real-time systems. The preferred list of each node is constructed in such a way that each node will be selected as the kth preferred node by one and only one other node. Such lists are proven to allow the tasks to be evenly distributed in a system. However, the presence of faulty nodes will destroy the original structure of a preferred list if the faulty nodes are simply skipped in the preferred list. An algorithm is therefore proposed to modify each preferred list to retain its original features regardless of the number of faulty nodes in the system. The communication overhead introduced by this algorithm is shown to be minimal. Based on the modified preferred lists, a simple fault-tolerant mechanism is implemented. Each node is equipped with a backup queue which 'The work reported in this paper was supported in part by the Office of Naval Research under contract N0001485-K-0122, and the NSF under grant DMC-8721492. Any opinions, findings, and recommendations expressed in this publication are those of the authors and do not necessarily reflect the view of the funding agencies. stores and updates the arriving/completing tasks at its most preferrecd node. Whenever a node becomes faulty, its m.ost preferred node will treat the tasks in the baxkup queue as externally axriving tasks. Our simulation results show that this approach, despite of the simplicity, can reduce the number of task losses dramatically, as compared to the approaches without any faulttolerant mechanism.","PeriodicalId":204431,"journal":{"name":"Proceedings of the Fifth Distributed Memory Computing Conference, 1990.","volume":"337 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115882207","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
Proceedings of the Fifth Distributed Memory Computing Conference, 1990.
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1