首页 > 最新文献

Proceedings of the Fifth Distributed Memory Computing Conference, 1990.最新文献

英文 中文
An Approach to Compiling Single-point Iterative Programs for Distributed Memory Computers 分布式存储计算机单点迭代程序的编译方法
Pub Date : 1990-04-08 DOI: 10.1109/DMCC.1990.556313
D. Socha
This paper proposes a scheme for compiling an important class of iterative algorithms into efficient code for distributed memory computers. The programmer provides a description of the problem in Spot: a data parallel SIMD language that uses iterations as the unit of synchronization and is based on grids of data points. The data parallel description is in terms of a single point of the data space, with implicit communication semantics, and a set of numerical boundary conditions. The compiler eliminates the need for multi-tasking by ‘(expanding” the single-point code into multiplepoint code that executes over rectangular regions of points. Using rectangle intersection and difference operations on these regions allows the compiler to automatically insert the required communication calls and to hide communication latency by overlapping comput at ion and communication. The multiple-point code may be specialized, at compile-time, to the size and shape of different allocations, or it may use table-driven for-loops to adapt, at run-time, to the shape and size of the allocations. We show how to generalize this strategy to produce code for the near-rectangular shaped allocations required for balanced partitionings of rectangular arrays.
本文提出了一种将一类重要的迭代算法编译成分布式存储计算机的高效代码的方案。程序员在Spot中提供了对问题的描述:一种数据并行SIMD语言,它使用迭代作为同步单元,并基于数据点网格。数据并行描述是根据数据空间的单点、隐式通信语义和一组数值边界条件进行的。编译器通过将单点代码“扩展”为在矩形点区域上执行的多点代码,消除了对多任务的需要。在这些区域上使用矩形相交和差分操作,可以使编译器自动插入所需的通信调用,并通过重叠计算和通信来隐藏通信延迟。多点代码可以在编译时专门用于不同分配的大小和形状,也可以使用表驱动的for循环在运行时适应分配的形状和大小。我们将展示如何推广此策略,以生成用于矩形数组均衡分区所需的近矩形分配的代码。
{"title":"An Approach to Compiling Single-point Iterative Programs for Distributed Memory Computers","authors":"D. Socha","doi":"10.1109/DMCC.1990.556313","DOIUrl":"https://doi.org/10.1109/DMCC.1990.556313","url":null,"abstract":"This paper proposes a scheme for compiling an important class of iterative algorithms into efficient code for distributed memory computers. The programmer provides a description of the problem in Spot: a data parallel SIMD language that uses iterations as the unit of synchronization and is based on grids of data points. The data parallel description is in terms of a single point of the data space, with implicit communication semantics, and a set of numerical boundary conditions. The compiler eliminates the need for multi-tasking by ‘(expanding” the single-point code into multiplepoint code that executes over rectangular regions of points. Using rectangle intersection and difference operations on these regions allows the compiler to automatically insert the required communication calls and to hide communication latency by overlapping comput at ion and communication. The multiple-point code may be specialized, at compile-time, to the size and shape of different allocations, or it may use table-driven for-loops to adapt, at run-time, to the shape and size of the allocations. We show how to generalize this strategy to produce code for the near-rectangular shaped allocations required for balanced partitionings of rectangular arrays.","PeriodicalId":204431,"journal":{"name":"Proceedings of the Fifth Distributed Memory Computing Conference, 1990.","volume":"87 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126014590","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Mapping Data to Processors in Distributed Memory Computations 分布式内存计算中数据到处理器的映射
Pub Date : 1990-04-08 DOI: 10.1109/DMCC.1990.556295
M. Rosing, R. P. Weaver
Abstract : The authors present a structured scheme for allowing a programmer to specify the mapping of data to distributed memory multiprocessors. This scheme lets the programmer specify information about communication patterns as well as information about distributing data structures onto processors (including partitioning with replication). This mapping scheme allows the user to map arrays of data to arrays of processors. The user specifies how each axis of the data structure is mapped onto an axis of the processor structure. This mapping may either be one to one or one to many depending on the parallelism, load balancing, and communication requirements. The authors discuss the basics of how this scheme is implemented in the DINO language, the areas in which it has worked well, the few areas in which there were significant problems, and some ideas for future improvements.
摘要:作者提出了一种结构化方案,允许程序员指定数据到分布式内存多处理器的映射。该方案允许程序员指定有关通信模式的信息以及有关将数据结构分布到处理器上的信息(包括带有复制的分区)。这种映射方案允许用户将数据数组映射到处理器数组。用户指定如何将数据结构的每个轴映射到处理器结构的轴上。这种映射可以是一对一的,也可以是一对多的,这取决于并行性、负载平衡和通信需求。作者讨论了该方案如何在DINO语言中实现的基础知识,它工作良好的领域,存在重大问题的少数领域,以及未来改进的一些想法。
{"title":"Mapping Data to Processors in Distributed Memory Computations","authors":"M. Rosing, R. P. Weaver","doi":"10.1109/DMCC.1990.556295","DOIUrl":"https://doi.org/10.1109/DMCC.1990.556295","url":null,"abstract":"Abstract : The authors present a structured scheme for allowing a programmer to specify the mapping of data to distributed memory multiprocessors. This scheme lets the programmer specify information about communication patterns as well as information about distributing data structures onto processors (including partitioning with replication). This mapping scheme allows the user to map arrays of data to arrays of processors. The user specifies how each axis of the data structure is mapped onto an axis of the processor structure. This mapping may either be one to one or one to many depending on the parallelism, load balancing, and communication requirements. The authors discuss the basics of how this scheme is implemented in the DINO language, the areas in which it has worked well, the few areas in which there were significant problems, and some ideas for future improvements.","PeriodicalId":204431,"journal":{"name":"Proceedings of the Fifth Distributed Memory Computing Conference, 1990.","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127401473","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
A Re-Configurable Reduced-Bus Multiprocessor Interconnection Network 一种可重新配置的减少总线的多处理器互连网络
Pub Date : 1990-04-08 DOI: 10.1109/DMCC.1990.556274
T. Ramesh, S. Ganesan
Multiple-bus multiprocessor interconnection networks are still considered as a cost effective and easily expandable processor-memory interconnection. But, a fully connected multiple bus network requires all busses to be connected to each processor and memory module thus increasing the physical connection and the bus load on each memo y. Reduced-bus connections with different connection topology such as rhombus, trapezoidal etc., were presented in [l]. In this paper a general single network topology has been presented that can be reconfigured to any one of the reduced-bus connection schemes. The re-configurability is achieved through the arbitration of combinations of simple link switches in a ring structure. The motive behind developing this reconfigurable structure is to offer a flexibility in matching the reduced-bus connection schemes to structure of parallel algorithms. The paper presents a mapping scheme to arbitrate the link switches for various connection pattems. Also, a comparison of effective memoy bandwidth of each connection scheme is shown. Expandability of the system to larger sizes are addressed.
多总线多处理器互连网络仍然被认为是一种经济有效且易于扩展的处理器-存储器互连网络。但是,一个完全连接的多总线网络要求所有总线都连接到每个处理器和存储器模块,从而增加了物理连接和每个备忘录上的总线负载。[1]提出了不同连接拓扑(如菱形、梯形等)的减少总线连接。本文提出了一种通用的单一网络拓扑结构,可以重新配置为任何一种减少总线连接方案。可重构性是通过环形结构中简单链路交换机组合的仲裁来实现的。开发这种可重构结构的动机是在将减少总线连接方案与并行算法结构相匹配方面提供灵活性。本文提出了一种映射方案来仲裁各种连接模式下的链路开关。此外,还比较了各种连接方案的有效内存带宽。系统的可扩展性,以更大的规模解决。
{"title":"A Re-Configurable Reduced-Bus Multiprocessor Interconnection Network","authors":"T. Ramesh, S. Ganesan","doi":"10.1109/DMCC.1990.556274","DOIUrl":"https://doi.org/10.1109/DMCC.1990.556274","url":null,"abstract":"Multiple-bus multiprocessor interconnection networks are still considered as a cost effective and easily expandable processor-memory interconnection. But, a fully connected multiple bus network requires all busses to be connected to each processor and memory module thus increasing the physical connection and the bus load on each memo y. Reduced-bus connections with different connection topology such as rhombus, trapezoidal etc., were presented in [l]. In this paper a general single network topology has been presented that can be reconfigured to any one of the reduced-bus connection schemes. The re-configurability is achieved through the arbitration of combinations of simple link switches in a ring structure. The motive behind developing this reconfigurable structure is to offer a flexibility in matching the reduced-bus connection schemes to structure of parallel algorithms. The paper presents a mapping scheme to arbitrate the link switches for various connection pattems. Also, a comparison of effective memoy bandwidth of each connection scheme is shown. Expandability of the system to larger sizes are addressed.","PeriodicalId":204431,"journal":{"name":"Proceedings of the Fifth Distributed Memory Computing Conference, 1990.","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132632523","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Dynamic Load Balancing in a Concurrent Plasma PIC Code on the JPL/Caltech Mark III Hypercube JPL/Caltech Mark III Hypercube上并发等离子体PIC代码的动态负载平衡
Pub Date : 1990-04-08 DOI: 10.1109/DMCC.1990.556302
P. Liewer, E. W. Leaver, V. Decyk, J. Dawson
Dynamic load balancing has been implemented in a concurrent one-dimensional electromagnetic plasma particle-in-cell (PIC) simulation code using a method which adds very little overhead to the parallel code. In PIC codes, the orbits of many interacting plasma electrons and ions are followed as an initial value problem as the particles move in electromagnetic fields calculated self-consistently from the particle motions. The code was implemented using the GCPIC algorithm in which the particles are divided among processors by partitioning the spatial domain of the simulation. The problem is load-balanced by partitioning the spatial domain so that each partition has approximately the same number of particles. During the simulation, the partitions are dynamically recreated as the spatial distribution of the particles changes in order to maintain processor load balance.
在一维电磁等离子体粒子池(PIC)仿真代码中实现了动态负载平衡,该方法对并行代码增加的开销很小。在PIC代码中,当粒子在自一致计算的电磁场中运动时,许多相互作用的等离子体电子和离子的轨道作为初始值问题被遵循。该代码使用GCPIC算法实现,其中粒子通过划分模拟的空间域在处理器之间进行划分。该问题通过划分空间域来实现负载平衡,以便每个分区具有大约相同数量的粒子。在模拟过程中,随着粒子空间分布的变化,动态地重新创建分区,以保持处理器负载平衡。
{"title":"Dynamic Load Balancing in a Concurrent Plasma PIC Code on the JPL/Caltech Mark III Hypercube","authors":"P. Liewer, E. W. Leaver, V. Decyk, J. Dawson","doi":"10.1109/DMCC.1990.556302","DOIUrl":"https://doi.org/10.1109/DMCC.1990.556302","url":null,"abstract":"Dynamic load balancing has been implemented in a concurrent one-dimensional electromagnetic plasma particle-in-cell (PIC) simulation code using a method which adds very little overhead to the parallel code. In PIC codes, the orbits of many interacting plasma electrons and ions are followed as an initial value problem as the particles move in electromagnetic fields calculated self-consistently from the particle motions. The code was implemented using the GCPIC algorithm in which the particles are divided among processors by partitioning the spatial domain of the simulation. The problem is load-balanced by partitioning the spatial domain so that each partition has approximately the same number of particles. During the simulation, the partitions are dynamically recreated as the spatial distribution of the particles changes in order to maintain processor load balance.","PeriodicalId":204431,"journal":{"name":"Proceedings of the Fifth Distributed Memory Computing Conference, 1990.","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130498587","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
The 600 Megaflops Performance of the QCD Code on the Mark IIIfp Hypercube Mark iii ifp超立方体上QCD代码的600兆次浮点运算性能
Pub Date : 1990-04-08 DOI: 10.1109/DMCC.1990.556389
H. Ding
{"title":"The 600 Megaflops Performance of the QCD Code on the Mark IIIfp Hypercube","authors":"H. Ding","doi":"10.1109/DMCC.1990.556389","DOIUrl":"https://doi.org/10.1109/DMCC.1990.556389","url":null,"abstract":"","PeriodicalId":204431,"journal":{"name":"Proceedings of the Fifth Distributed Memory Computing Conference, 1990.","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117177565","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Local Search Variants for Hypercube Embedding 超立方体嵌入的局部搜索变体
Pub Date : 1990-04-08 DOI: 10.1109/DMCC.1990.556399
Woei-kae Chen, Matthias F. Stallmann
The hypercube embedding problem, a restricted ver- sion of the general mapping problem, is the problem of mapping a set of communicating processes to a hy- percube multiprocessor. The goal is to find a map- ping that minimizes the average length of the paths between communicating processes. Iterative improve- ment heuristics for hypercube embedding, including a local search, a Kernighan-Lin, and a simulated an- nealing, are evaluated under different options includ- ing neighborhoods (all-swaps versus cube-neighbors), initial solutions (random versus greedy), and enhance- ments on terminating conditions (flat moves and up- hill moves). By varying these options we obtain a wide range of tradeoffs between execution time and solution quality.
超立方体嵌入问题是一般映射问题的一个限制版本,它是将一组通信进程映射到一个超立方体多处理器的问题。目标是找到一种映射,使通信进程之间的平均路径长度最小化。超立方体嵌入的迭代改进启发式,包括局部搜索、Kernighan-Lin和模拟逼近,在不同的选项下进行评估,包括邻域(全交换与立方体邻居)、初始解(随机与贪婪)和终止条件的增强(平移和上坡移动)。通过改变这些选项,我们可以在执行时间和解决方案质量之间获得广泛的折衷。
{"title":"Local Search Variants for Hypercube Embedding","authors":"Woei-kae Chen, Matthias F. Stallmann","doi":"10.1109/DMCC.1990.556399","DOIUrl":"https://doi.org/10.1109/DMCC.1990.556399","url":null,"abstract":"The hypercube embedding problem, a restricted ver- sion of the general mapping problem, is the problem of mapping a set of communicating processes to a hy- percube multiprocessor. The goal is to find a map- ping that minimizes the average length of the paths between communicating processes. Iterative improve- ment heuristics for hypercube embedding, including a local search, a Kernighan-Lin, and a simulated an- nealing, are evaluated under different options includ- ing neighborhoods (all-swaps versus cube-neighbors), initial solutions (random versus greedy), and enhance- ments on terminating conditions (flat moves and up- hill moves). By varying these options we obtain a wide range of tradeoffs between execution time and solution quality.","PeriodicalId":204431,"journal":{"name":"Proceedings of the Fifth Distributed Memory Computing Conference, 1990.","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128421556","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Reducing Inner Product Computation in the Parallel One-Sided Jacobi Algorithm 减少平行单侧Jacobi算法的内积计算
Pub Date : 1990-04-08 DOI: 10.1109/DMCC.1990.555398
C. Romine, K. Sigmon
{"title":"Reducing Inner Product Computation in the Parallel One-Sided Jacobi Algorithm","authors":"C. Romine, K. Sigmon","doi":"10.1109/DMCC.1990.555398","DOIUrl":"https://doi.org/10.1109/DMCC.1990.555398","url":null,"abstract":"","PeriodicalId":204431,"journal":{"name":"Proceedings of the Fifth Distributed Memory Computing Conference, 1990.","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"113977152","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient Serial and Parallel Subcube Recognition in Hypercubes 超立方体中高效的串行和并行子立方体识别
Pub Date : 1990-04-08 DOI: 10.1109/DMCC.1990.555363
S. Al-Bassam, H. El-Rewini, B. Bose, T. Lewis
We develop an efficient subcube recognition algorithm that recognizes all the possible subcubes. The algorithm is based on exploiting more subcubes at different levels of the buddy tree. In exploiting the different levels, the algorithm checks any subcube at most once. Moreover, many unavailable subcubes are not considered as candidates and hence not checked for availability. This makes the algorithm fast in recognizing the subcubes. The number of recognized subcubes, for different subcube sizes, can be easily adjusted by restricting the search level down the buddy tree. The previous known algorithms become a special case of this general approach. When one level is searched, this algorithm perfoms as the original buddy system. When two levels are searched, it will recognized the Same subcubes as the ones in [4] with a faster speed. When all the levels are searched, a complete subcube recognition is obtained. In a multi-processing system, each processor can execute this algorithm on a different tree. Using a given number of processors in a multi-processing system, we give a method of constructing the trees that maximizes the overall number of recognized subcubes. Finally, we introduce an allocation method "best fit" that reduces hypercube fragmentation. Simulation results and performance comparisons between this method and the traditional "first fit" are presented.
我们开发了一种有效的子立方体识别算法,可以识别所有可能的子立方体。该算法基于在伙伴树的不同层次上挖掘更多的子数据集。在利用不同的层次时,该算法最多检查一次任何子立方体。此外,许多不可用的子数据集不被视为候选数据集,因此不检查可用性。这使得该算法能够快速识别子数据集。对于不同大小的子多维数据集,可以通过限制好友树的搜索级别来轻松调整可识别的子多维数据集的数量。以前已知的算法成为这种通用方法的特例。当搜索一个关卡时,该算法执行原始的伙伴系统。当搜索两个级别时,它将以更快的速度识别与[4]中相同的子数据集。当搜索所有级别时,获得完整的子立方体识别。在多处理系统中,每个处理器可以在不同的树上执行该算法。在多处理系统中使用给定数量的处理器,我们给出了一种构造树的方法,该树可以最大限度地增加可识别子数据集的总数。最后,我们引入了一种减少超立方体碎片的“最佳拟合”分配方法。给出了该方法与传统的“首次拟合”方法的仿真结果和性能比较。
{"title":"Efficient Serial and Parallel Subcube Recognition in Hypercubes","authors":"S. Al-Bassam, H. El-Rewini, B. Bose, T. Lewis","doi":"10.1109/DMCC.1990.555363","DOIUrl":"https://doi.org/10.1109/DMCC.1990.555363","url":null,"abstract":"We develop an efficient subcube recognition algorithm that recognizes all the possible subcubes. The algorithm is based on exploiting more subcubes at different levels of the buddy tree. In exploiting the different levels, the algorithm checks any subcube at most once. Moreover, many unavailable subcubes are not considered as candidates and hence not checked for availability. This makes the algorithm fast in recognizing the subcubes. The number of recognized subcubes, for different subcube sizes, can be easily adjusted by restricting the search level down the buddy tree. The previous known algorithms become a special case of this general approach. When one level is searched, this algorithm perfoms as the original buddy system. When two levels are searched, it will recognized the Same subcubes as the ones in [4] with a faster speed. When all the levels are searched, a complete subcube recognition is obtained. In a multi-processing system, each processor can execute this algorithm on a different tree. Using a given number of processors in a multi-processing system, we give a method of constructing the trees that maximizes the overall number of recognized subcubes. Finally, we introduce an allocation method \"best fit\" that reduces hypercube fragmentation. Simulation results and performance comparisons between this method and the traditional \"first fit\" are presented.","PeriodicalId":204431,"journal":{"name":"Proceedings of the Fifth Distributed Memory Computing Conference, 1990.","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123955406","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Hypercubes for Critical Space Flight Command Operations 关键空间飞行指挥操作的超立方体
Pub Date : 1990-04-08 DOI: 10.1109/DMCC.1990.555355
J. Horvath, T. Tang, L. P. Perry, R. Cole, D.B. OIster, J. Zipse
Controlling interplanetary spacecraft and planning their activities, as currently practiced, requires massive amounts of computer time and personnel. To improve this situation, it is desired to use advanced computing to speed up and automate the commanding process. Several design and prototype efforts have been underway at JPL to understand the appropriate roles for concurrent processors in future interplanetary spacecraft operations. Here we report on an effort to identify likely candidates for parallelism among existing software systems that both generate commands to be sent to the spacecraft and simulate what the spacecraft will do with these commands when it receives them. We also describe promising results from efforts to create parallel prototypes of representative portions of these software systems on the JPL/Caltech Mark 111 hypercube.
控制行星际航天器并规划其活动,如目前所实行的,需要大量的计算机时间和人员。为了改善这种情况,需要使用先进的计算来加速和自动化指挥过程。喷气推进实验室正在进行一些设计和原型工作,以了解并发处理器在未来星际航天器操作中的适当作用。在这里,我们报告了一项努力,以确定现有软件系统中可能的并行候选,这些软件系统既生成要发送给航天器的命令,又模拟航天器在接收这些命令时将如何处理这些命令。我们还描述了在JPL/Caltech Mark 111超立方体上创建这些软件系统的代表性部分的并行原型的有希望的结果。
{"title":"Hypercubes for Critical Space Flight Command Operations","authors":"J. Horvath, T. Tang, L. P. Perry, R. Cole, D.B. OIster, J. Zipse","doi":"10.1109/DMCC.1990.555355","DOIUrl":"https://doi.org/10.1109/DMCC.1990.555355","url":null,"abstract":"Controlling interplanetary spacecraft and planning their activities, as currently practiced, requires massive amounts of computer time and personnel. To improve this situation, it is desired to use advanced computing to speed up and automate the commanding process. Several design and prototype efforts have been underway at JPL to understand the appropriate roles for concurrent processors in future interplanetary spacecraft operations. Here we report on an effort to identify likely candidates for parallelism among existing software systems that both generate commands to be sent to the spacecraft and simulate what the spacecraft will do with these commands when it receives them. We also describe promising results from efforts to create parallel prototypes of representative portions of these software systems on the JPL/Caltech Mark 111 hypercube.","PeriodicalId":204431,"journal":{"name":"Proceedings of the Fifth Distributed Memory Computing Conference, 1990.","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127664534","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Load Sharing In Hypercube Multicomputers In The Presence Of Node Failures 节点故障情况下超立方体多计算机的负载分担
Pub Date : 1990-04-08 DOI: 10.1109/DMCC.1990.556410
Yi-Chieh Chang, K. Shin
This paper discusses and analyzes two load sharing (LS) issues: adjusting preferred lists and implementing a fault-tolerant mechanism in the presence of node failures. In an early paper, we have proposed to order the nodes in each node's proximity into a preferred list for the purpose of load sharing in distributed real-time systems. The preferred list of each node is constructed in such a way that each node will be selected as the kth preferred node by one and only one other node. Such lists are proven to allow the tasks to be evenly distributed in a system. However, the presence of faulty nodes will destroy the original structure of a preferred list if the faulty nodes are simply skipped in the preferred list. An algorithm is therefore proposed to modify each preferred list to retain its original features regardless of the number of faulty nodes in the system. The communication overhead introduced by this algorithm is shown to be minimal. Based on the modified preferred lists, a simple fault-tolerant mechanism is implemented. Each node is equipped with a backup queue which 'The work reported in this paper was supported in part by the Office of Naval Research under contract N0001485-K-0122, and the NSF under grant DMC-8721492. Any opinions, findings, and recommendations expressed in this publication are those of the authors and do not necessarily reflect the view of the funding agencies. stores and updates the arriving/completing tasks at its most preferrecd node. Whenever a node becomes faulty, its m.ost preferred node will treat the tasks in the baxkup queue as externally axriving tasks. Our simulation results show that this approach, despite of the simplicity, can reduce the number of task losses dramatically, as compared to the approaches without any faulttolerant mechanism.
本文讨论并分析了负载共享(LS)的两个问题:调整首选列表和实现节点故障时的容错机制。在早期的一篇论文中,我们提出了在分布式实时系统中,为了负载共享的目的,将每个节点邻近的节点排序到一个首选列表中。构建每个节点的首选列表的方式是,每个节点将被一个且仅一个其他节点选择为第k个首选节点。这样的列表被证明允许任务在系统中均匀分布。但是,如果在首选列表中跳过故障节点,则故障节点的存在将破坏首选列表的原始结构。因此,提出了一种算法,无论系统中故障节点的数量如何,都可以修改每个首选列表以保留其原始特征。该算法带来的通信开销是最小的。基于修改后的首选列表,实现了简单的容错机制。每个节点都配备了一个备份队列,本文报告的工作部分由海军研究办公室根据合同N0001485-K-0122和NSF根据拨款DMC-8721492提供支持。本出版物中表达的任何观点、发现和建议均为作者的观点,并不一定反映资助机构的观点。在其最优选的节点上存储和更新到达/完成的任务。当一个节点发生故障时,它的最优节点将把备份队列中的任务视为外部到达的任务。我们的仿真结果表明,尽管这种方法简单,但与没有容错机制的方法相比,可以显着减少任务损失的数量。
{"title":"Load Sharing In Hypercube Multicomputers In The Presence Of Node Failures","authors":"Yi-Chieh Chang, K. Shin","doi":"10.1109/DMCC.1990.556410","DOIUrl":"https://doi.org/10.1109/DMCC.1990.556410","url":null,"abstract":"This paper discusses and analyzes two load sharing (LS) issues: adjusting preferred lists and implementing a fault-tolerant mechanism in the presence of node failures. In an early paper, we have proposed to order the nodes in each node's proximity into a preferred list for the purpose of load sharing in distributed real-time systems. The preferred list of each node is constructed in such a way that each node will be selected as the kth preferred node by one and only one other node. Such lists are proven to allow the tasks to be evenly distributed in a system. However, the presence of faulty nodes will destroy the original structure of a preferred list if the faulty nodes are simply skipped in the preferred list. An algorithm is therefore proposed to modify each preferred list to retain its original features regardless of the number of faulty nodes in the system. The communication overhead introduced by this algorithm is shown to be minimal. Based on the modified preferred lists, a simple fault-tolerant mechanism is implemented. Each node is equipped with a backup queue which 'The work reported in this paper was supported in part by the Office of Naval Research under contract N0001485-K-0122, and the NSF under grant DMC-8721492. Any opinions, findings, and recommendations expressed in this publication are those of the authors and do not necessarily reflect the view of the funding agencies. stores and updates the arriving/completing tasks at its most preferrecd node. Whenever a node becomes faulty, its m.ost preferred node will treat the tasks in the baxkup queue as externally axriving tasks. Our simulation results show that this approach, despite of the simplicity, can reduce the number of task losses dramatically, as compared to the approaches without any faulttolerant mechanism.","PeriodicalId":204431,"journal":{"name":"Proceedings of the Fifth Distributed Memory Computing Conference, 1990.","volume":"337 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115882207","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
Proceedings of the Fifth Distributed Memory Computing Conference, 1990.
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1