首页 > 最新文献

[1990 Proceedings] The Third Symposium on the Frontiers of Massively Parallel Computation最新文献

英文 中文
Divacon: a parallel language for scientific computing based on divide-and-conquer Divacon:一种基于分治法的科学计算并行语言
Z. G. Mou
An overview of the language, covering Divacon primitives and simple programming constructs that are referred to as functional forms, is given. Two divide-and-conquer programming constructs are discussed. Divacon style programming is demonstrated for a number of scientific applications. Some interesting equivalences and transformations between Divacon programs are examined. Implementation and performance are briefly considered.<>
给出了该语言的概述,包括Divacon原语和称为函数形式的简单编程结构。讨论了两种分而治之的编程结构。Divacon风格的编程在许多科学应用中得到了演示。研究了Divacon程序之间的一些有趣的等价和转换。简要地考虑了实现和性能
{"title":"Divacon: a parallel language for scientific computing based on divide-and-conquer","authors":"Z. G. Mou","doi":"10.1109/FMPC.1990.89496","DOIUrl":"https://doi.org/10.1109/FMPC.1990.89496","url":null,"abstract":"An overview of the language, covering Divacon primitives and simple programming constructs that are referred to as functional forms, is given. Two divide-and-conquer programming constructs are discussed. Divacon style programming is demonstrated for a number of scientific applications. Some interesting equivalences and transformations between Divacon programs are examined. Implementation and performance are briefly considered.<<ETX>>","PeriodicalId":193332,"journal":{"name":"[1990 Proceedings] The Third Symposium on the Frontiers of Massively Parallel Computation","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129678682","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
A distributed backpropagation algorithm of neural networks on distributed-memory multiprocessors 分布式存储多处理器上神经网络的分布式反向传播算法
H. Yoon, J.H. Nang, S. Maeng
A distributed backpropagation algorithm for a fully connected multilayered neural network on a distributed-memory multiprocessor system is presented. The neurons on each layer are partitioned into p disjoint sets, and each set is mapped on a processor of a p-processor system. The algorithm, the communication pattern among the processors, and their time/space complexities are investigated, and the theoretical upper bound on speedup is obtained. The experimental speedup obtained with the algorithm on a ring of 32 transputers, which confirms the model and analysis, is reported. It is found that the choice of processor interconnection topology does not influence the speedup ratio.<>
提出了一种基于分布式存储多处理器系统的全连接多层神经网络的分布式反向传播算法。每一层的神经元被划分为p个不相交的集合,每个集合被映射到p处理器系统的一个处理器上。研究了该算法、处理器间的通信模式以及它们的时间/空间复杂度,得到了加速的理论上界。在32个传感器环上进行了加速实验,验证了模型和分析的正确性。研究发现,处理器互连拓扑的选择对加速比没有影响。
{"title":"A distributed backpropagation algorithm of neural networks on distributed-memory multiprocessors","authors":"H. Yoon, J.H. Nang, S. Maeng","doi":"10.1109/FMPC.1990.89482","DOIUrl":"https://doi.org/10.1109/FMPC.1990.89482","url":null,"abstract":"A distributed backpropagation algorithm for a fully connected multilayered neural network on a distributed-memory multiprocessor system is presented. The neurons on each layer are partitioned into p disjoint sets, and each set is mapped on a processor of a p-processor system. The algorithm, the communication pattern among the processors, and their time/space complexities are investigated, and the theoretical upper bound on speedup is obtained. The experimental speedup obtained with the algorithm on a ring of 32 transputers, which confirms the model and analysis, is reported. It is found that the choice of processor interconnection topology does not influence the speedup ratio.<<ETX>>","PeriodicalId":193332,"journal":{"name":"[1990 Proceedings] The Third Symposium on the Frontiers of Massively Parallel Computation","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121879167","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Deterministic PRAM simulation with constant memory blow-up and no time-stamps 具有恒定内存爆炸和无时间戳的确定性PRAM仿真
Y. Aumann, A. Schuster
A scheme for deterministic simulation of a parallel random-access machine (PRAM) on a module parallel computer or on bounded-degree networks is described. The scheme requires only a constant memory blowup, thus achieving better memory utilization than previously known approaches. The method does not need time stamps, which were a basic element of all previous schemes. The improvements are achieved by adopting error-correcting-code techniques. Several coding methods are considered, tradeoffs between memory utilization, run time, and the size of the PRAM shared memory are derived.<>
提出了一种在模块并行计算机或有界度网络上对并行随机存取机(PRAM)进行确定性仿真的方案。该方案只需要不断增加内存,因此比以前已知的方法获得更好的内存利用率。该方法不需要时间戳,而时间戳是以前所有方案的基本元素。改进是通过采用纠错码技术实现的。考虑了几种编码方法,得出了内存利用率、运行时间和PRAM共享内存大小之间的权衡
{"title":"Deterministic PRAM simulation with constant memory blow-up and no time-stamps","authors":"Y. Aumann, A. Schuster","doi":"10.1109/FMPC.1990.89431","DOIUrl":"https://doi.org/10.1109/FMPC.1990.89431","url":null,"abstract":"A scheme for deterministic simulation of a parallel random-access machine (PRAM) on a module parallel computer or on bounded-degree networks is described. The scheme requires only a constant memory blowup, thus achieving better memory utilization than previously known approaches. The method does not need time stamps, which were a basic element of all previous schemes. The improvements are achieved by adopting error-correcting-code techniques. Several coding methods are considered, tradeoffs between memory utilization, run time, and the size of the PRAM shared memory are derived.<<ETX>>","PeriodicalId":193332,"journal":{"name":"[1990 Proceedings] The Third Symposium on the Frontiers of Massively Parallel Computation","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126148467","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Parallel optimization of stack filters 堆栈滤波器的并行优化
K. Rao, K. Efe, C. H. Chu
An open problem associated with designing stack filters is finding the optimum configuration for a given noise type and the signal characteristics which need to be preserved. This problem is modeled here as a combinatorial search problem. Efficient search methods that can be easily implemented on any massively parallel computer were developed and tested in two parallel computing environments. The first is the Connection Machine, and the second is a hypercube-connected MIMD (multiple-instruction-stream, multiple-data-stream) machine simulated using Cosmic C. The performance of the filters found by the algorithms developed were excellent in comparison with the performance of the median filter. The efficiency of the algorithms clearly demonstrates the potential of using them for adaptive filtering. The algorithms can be implemented on any type of parallel computer.<>
设计堆叠滤波器的一个开放问题是,在给定噪声类型和需要保留的信号特性下,找到最佳配置。这个问题在这里被建模为一个组合搜索问题。在两个并行计算环境中开发并测试了可以在任何大规模并行计算机上轻松实现的高效搜索方法。第一个是连接机器,第二个是使用Cosmic c模拟的超立方体连接的MIMD(多指令流,多数据流)机器。与中值滤波器的性能相比,所开发算法发现的滤波器的性能非常出色。这些算法的效率清楚地表明了将它们用于自适应滤波的潜力。该算法可在任何类型的并行计算机上实现。
{"title":"Parallel optimization of stack filters","authors":"K. Rao, K. Efe, C. H. Chu","doi":"10.1109/FMPC.1990.89505","DOIUrl":"https://doi.org/10.1109/FMPC.1990.89505","url":null,"abstract":"An open problem associated with designing stack filters is finding the optimum configuration for a given noise type and the signal characteristics which need to be preserved. This problem is modeled here as a combinatorial search problem. Efficient search methods that can be easily implemented on any massively parallel computer were developed and tested in two parallel computing environments. The first is the Connection Machine, and the second is a hypercube-connected MIMD (multiple-instruction-stream, multiple-data-stream) machine simulated using Cosmic C. The performance of the filters found by the algorithms developed were excellent in comparison with the performance of the median filter. The efficiency of the algorithms clearly demonstrates the potential of using them for adaptive filtering. The algorithms can be implemented on any type of parallel computer.<<ETX>>","PeriodicalId":193332,"journal":{"name":"[1990 Proceedings] The Third Symposium on the Frontiers of Massively Parallel Computation","volume":"90 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129453117","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Generalized supercube: an incrementally expandable interconnection network 广义超立方体:一种增量可扩展互联网络
Arunabha Sen, A. Sengupta, S. Bandyopadhyay
A class of incrementally expandable interconnection networks with high connectivity and low diameter is introduced for massively parallel and distributed processing. This class of networks can be constructed for any number of computing nodes, and the network size can easily be incremented without a major reconfiguration of the network. The connectivity and the diameter of the network are on the order of the logarithm of the number of nodes. It is shown that the connectivity of the network is equal to the minimum node degree. In this sense the connectivity is optimal. The routing algorithms for the network ar simple to implement.<>
针对大规模并行和分布式处理,提出了一类高连通性、低直径增量可扩展互连网络。这类网络可以为任意数量的计算节点构建,并且可以轻松地增加网络大小,而无需对网络进行重大重新配置。网络的连通性和直径是节点数的对数数量级。结果表明,网络的连通性等于最小节点度。从这个意义上说,连通性是最优的。该网络的路由算法易于实现。
{"title":"Generalized supercube: an incrementally expandable interconnection network","authors":"Arunabha Sen, A. Sengupta, S. Bandyopadhyay","doi":"10.1109/FMPC.1990.89488","DOIUrl":"https://doi.org/10.1109/FMPC.1990.89488","url":null,"abstract":"A class of incrementally expandable interconnection networks with high connectivity and low diameter is introduced for massively parallel and distributed processing. This class of networks can be constructed for any number of computing nodes, and the network size can easily be incremented without a major reconfiguration of the network. The connectivity and the diameter of the network are on the order of the logarithm of the number of nodes. It is shown that the connectivity of the network is equal to the minimum node degree. In this sense the connectivity is optimal. The routing algorithms for the network ar simple to implement.<<ETX>>","PeriodicalId":193332,"journal":{"name":"[1990 Proceedings] The Third Symposium on the Frontiers of Massively Parallel Computation","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121524972","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
Topological properties of banyan-hypercube networks 榕树-超立方体网络的拓扑性质
A. Youssef, B. Narahari
Topological properties of banyan-hypercubes are discussed, and a family of generalized banyan-hypercubes is defined. A banyan-hypercube, denoted BH(h, k, s), is constructed by taking the bottom h levels of a rectangular banyan of spread s and s/sup k/ nodes per level for s a power of two, and interconnecting the nodes at each level in a hypercube. BHs can be viewed as a scheme for interconnecting hypercubes while keeping most of the advantages of the latter. The definition of BHs is extended and generalized to allow the interconnection of an unlimited number of hypercubes and to allow any h successive levels of the banyan to interconnect hypercubes. This leads to better extendibility and flexibility in partitioning the BH. The diameter and average distance of the generalized BH are derived and are shown to provide an improvement over the hypercube for a wide range of h, k, and s values. Self-routing point-to-point and broadcasting algorithms are presented, and efficient embeddings of various networks on the BH are shown.<>
讨论了榕树超立方体的拓扑性质,并定义了一类广义榕树超立方体。一个榕树-超立方体,记作BH(h, k, s),是通过取分布为s的矩形榕树的底部h层和s/sup / k/节点的每层(s为2的幂),并将超立方体中每层的节点互连来构建的。BHs可以被看作是一种连接超立方体的方案,同时保留了后者的大部分优点。对BHs的定义进行了扩展和推广,以允许无限数量的超立方体互连,并允许任意h个连续的banyan层互连超立方体。这样可以在划分BH时获得更好的可扩展性和灵活性。推导了广义黑洞的直径和平均距离,并证明它们在大范围的h、k和s值上优于超立方体。提出了自路由点对点和广播算法,并展示了各种网络在BH上的有效嵌入。
{"title":"Topological properties of banyan-hypercube networks","authors":"A. Youssef, B. Narahari","doi":"10.1109/FMPC.1990.89478","DOIUrl":"https://doi.org/10.1109/FMPC.1990.89478","url":null,"abstract":"Topological properties of banyan-hypercubes are discussed, and a family of generalized banyan-hypercubes is defined. A banyan-hypercube, denoted BH(h, k, s), is constructed by taking the bottom h levels of a rectangular banyan of spread s and s/sup k/ nodes per level for s a power of two, and interconnecting the nodes at each level in a hypercube. BHs can be viewed as a scheme for interconnecting hypercubes while keeping most of the advantages of the latter. The definition of BHs is extended and generalized to allow the interconnection of an unlimited number of hypercubes and to allow any h successive levels of the banyan to interconnect hypercubes. This leads to better extendibility and flexibility in partitioning the BH. The diameter and average distance of the generalized BH are derived and are shown to provide an improvement over the hypercube for a wide range of h, k, and s values. Self-routing point-to-point and broadcasting algorithms are presented, and efficient embeddings of various networks on the BH are shown.<<ETX>>","PeriodicalId":193332,"journal":{"name":"[1990 Proceedings] The Third Symposium on the Frontiers of Massively Parallel Computation","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114358324","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Parallel relational operations based on clustered surrogate files 基于集群代理文件的并行关系操作
S. M. Chung
In the context of very large databases, the main problem is how to process relational operations in the minimum amount of time to satisfy user queries. To speed up the relational operations on very large databases, parallel processing is essential. A reasonable indexing scheme for parallel processing systems is the concatenated code word (CCW) surrogate file, which is small in size and requires simple maintenance. Since interrelated relational operations can be performed on the CCW surrogate files, considerable processing time can be saved by performing interrelated relational operations on CCW files before the large data files are accessed. CCW surrogate files can be satisfactorily mapped into parallel architectures because their structure is quite compact and regular. To speed up the relational operations based on CCW surrogate files. it is possible to cluster the CCW surrogate files. If a CCW surrogate file is clustered, only a subset of the surrogate file is searched for a relational operation. Clustered-CCW surrogate file and data file structures suitable for a parallel processing system are introduced. Parallel relational operation algorithms based on the clustered file structures are developed and evaluated.<>
在非常大的数据库环境中,主要问题是如何在最短的时间内处理关系操作以满足用户查询。为了加快大型数据库上的关系操作,并行处理是必不可少的。对于并行处理系统来说,一个合理的索引方案是连接码字(CCW)代理文件,它体积小,维护简单。由于可以对CCW代理文件执行相关的关系操作,因此在访问大数据文件之前,通过对CCW文件执行相关的关系操作,可以节省大量的处理时间。CCW代理文件可以令人满意地映射到并行体系结构中,因为它们的结构非常紧凑和规则。提高基于CCW代理文件的关系操作速度。可以对CCW代理文件进行聚类。如果对CCW代理文件进行集群化,则只搜索代理文件的一个子集来进行关系操作。介绍了适用于并行处理系统的集群ccw代理文件和数据文件结构。提出并评价了基于聚类文件结构的并行关系运算算法。
{"title":"Parallel relational operations based on clustered surrogate files","authors":"S. M. Chung","doi":"10.1109/FMPC.1990.89463","DOIUrl":"https://doi.org/10.1109/FMPC.1990.89463","url":null,"abstract":"In the context of very large databases, the main problem is how to process relational operations in the minimum amount of time to satisfy user queries. To speed up the relational operations on very large databases, parallel processing is essential. A reasonable indexing scheme for parallel processing systems is the concatenated code word (CCW) surrogate file, which is small in size and requires simple maintenance. Since interrelated relational operations can be performed on the CCW surrogate files, considerable processing time can be saved by performing interrelated relational operations on CCW files before the large data files are accessed. CCW surrogate files can be satisfactorily mapped into parallel architectures because their structure is quite compact and regular. To speed up the relational operations based on CCW surrogate files. it is possible to cluster the CCW surrogate files. If a CCW surrogate file is clustered, only a subset of the surrogate file is searched for a relational operation. Clustered-CCW surrogate file and data file structures suitable for a parallel processing system are introduced. Parallel relational operation algorithms based on the clustered file structures are developed and evaluated.<<ETX>>","PeriodicalId":193332,"journal":{"name":"[1990 Proceedings] The Third Symposium on the Frontiers of Massively Parallel Computation","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115341374","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Toward scalable algorithms for orthogonal shared-memory parallel computers 正交共享内存并行计算机的可扩展算法
I. Scherson, A. Mehra, J. Rexford
The problem of developing scalable and near-optimal algorithms for orthogonal shared-memory multiprocessing systems with a multidimensional access (MDA) memory array is considered. An orthogonal shared-memory system consists of 2/sup n/ processors and 2/sup m/ memory modules accessed in any one of m possible access modes. Data stored in memory modules are available to processors under a mapping rule that allows conflict-free data reads and writes for any given access mode. Scalable algorithms are presented for two well-known computational problems, namely, matrix multiplication and the fast Fourier transform (FFT). A complete analysis of the algorithms based on computational time and the access modes needed is also presented. The algorithms scale very well onto higher dimensional MDA architectures but are not always optimal. This reveals a tradeoff between the scalability of an algorithm and its optimality in the MDA computational model.<>
研究了具有多维访问(MDA)存储阵列的正交共享内存多处理系统的可扩展和近最优算法开发问题。一个正交共享内存系统由2/sup n/个处理器和2/sup m/个内存模块组成,这些模块以m种可能的访问方式中的任意一种访问。存储在内存模块中的数据在映射规则下可供处理器使用,该规则允许对任何给定的访问模式进行无冲突的数据读取和写入。针对矩阵乘法和快速傅里叶变换(FFT)这两个众所周知的计算问题,提出了可扩展算法。对基于计算时间和所需访问方式的算法进行了全面的分析。这些算法可以很好地扩展到更高维度的MDA体系结构上,但并不总是最优的。这揭示了在MDA计算模型中算法的可伸缩性和最优性之间的权衡。
{"title":"Toward scalable algorithms for orthogonal shared-memory parallel computers","authors":"I. Scherson, A. Mehra, J. Rexford","doi":"10.1109/FMPC.1990.89430","DOIUrl":"https://doi.org/10.1109/FMPC.1990.89430","url":null,"abstract":"The problem of developing scalable and near-optimal algorithms for orthogonal shared-memory multiprocessing systems with a multidimensional access (MDA) memory array is considered. An orthogonal shared-memory system consists of 2/sup n/ processors and 2/sup m/ memory modules accessed in any one of m possible access modes. Data stored in memory modules are available to processors under a mapping rule that allows conflict-free data reads and writes for any given access mode. Scalable algorithms are presented for two well-known computational problems, namely, matrix multiplication and the fast Fourier transform (FFT). A complete analysis of the algorithms based on computational time and the access modes needed is also presented. The algorithms scale very well onto higher dimensional MDA architectures but are not always optimal. This reveals a tradeoff between the scalability of an algorithm and its optimality in the MDA computational model.<<ETX>>","PeriodicalId":193332,"journal":{"name":"[1990 Proceedings] The Third Symposium on the Frontiers of Massively Parallel Computation","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123400285","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PRA*: a memory-limited heuristic search procedure for the Connection Machine PRA*:一个内存有限的连接机启发式搜索过程
M. Evett, James A. Hendler, A. Mahanti, Dana S. Nau
A variant of A* search designed to run on the massively parallel SIMD (single-instruction-stream, multiple-data-steam) Connection Machine is described. The algorithm is designed to run in a limited memory; a retraction technique allows nodes with poor heuristic values to be removed from the open list until such time as they may need reexpansion if more promising paths fail. The algorithm, called PRA* (for parallel retraction A*), takes maximum advantage of the SIMD design of the Connection Machine and is guaranteed to return an optimal path when an admissible heuristic is used. Results comparing PRA* to R. Korf's IDA* (see Artif. Intell. J., vol.27, 1985) for the 15 puzzle show significantly fewer node expansions for PRA*.<>
描述了A* search的一个变体,该变体设计用于在大规模并行SIMD(单指令流,多数据流)连接机上运行。该算法被设计为在有限的内存中运行;收缩技术允许将启发式值较差的节点从打开列表中删除,直到如果更有希望的路径失败,它们可能需要重新展开。该算法称为PRA*(表示并行回缩A*),它最大限度地利用了连接机的SIMD设计,并保证在使用可接受的启发式时返回最优路径。PRA*与R. Korf的IDA*的比较结果(见Artif。智能。J., vol.27, 1985)对于15个谜题,PRA*.>的节点扩展显著减少
{"title":"PRA*: a memory-limited heuristic search procedure for the Connection Machine","authors":"M. Evett, James A. Hendler, A. Mahanti, Dana S. Nau","doi":"10.1109/FMPC.1990.89450","DOIUrl":"https://doi.org/10.1109/FMPC.1990.89450","url":null,"abstract":"A variant of A* search designed to run on the massively parallel SIMD (single-instruction-stream, multiple-data-steam) Connection Machine is described. The algorithm is designed to run in a limited memory; a retraction technique allows nodes with poor heuristic values to be removed from the open list until such time as they may need reexpansion if more promising paths fail. The algorithm, called PRA* (for parallel retraction A*), takes maximum advantage of the SIMD design of the Connection Machine and is guaranteed to return an optimal path when an admissible heuristic is used. Results comparing PRA* to R. Korf's IDA* (see Artif. Intell. J., vol.27, 1985) for the 15 puzzle show significantly fewer node expansions for PRA*.<<ETX>>","PeriodicalId":193332,"journal":{"name":"[1990 Proceedings] The Third Symposium on the Frontiers of Massively Parallel Computation","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128736284","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 42
The GPA machine: a generally partitionable MSIMD architecture GPA机器:一般可分区的MSIMD体系结构
T. Bridges
The GPA machine, a massively parallel, multiple single-instruction-stream-multiple-data-stream (MSIMD) system is described. Its distinguishing characteristics is the generality of its partitioning capabilities. Like the PASM system it can be dynamically reconfigured to operate as one or more independent SIMD machines. However, unlike PASM, the only constraint placed on partitioning is that an individual processing element is a member of at most one partition. This capability allows for reconfiguration based on the run-time status of dynamic data structures and for partitioning of disconnected and overlapping data structures. Significant speedups are expected from operating on data structures in place; copying of data to a newly configured partition is unnecessary. The GPA system consists of N processing-element/RAM pairs and an interconnection network providing access to and from P control processors or microcontrollers. With current technologies, values for N and P of 64K and 16, respectively, are feasible.<>
介绍了一种大规模并行、多单指令流多数据流(MSIMD)系统——GPA机。其显著特征是其分区能力的通用性。与PASM系统一样,它可以动态地重新配置为作为一个或多个独立的SIMD机器运行。然而,与PASM不同的是,对分区的唯一约束是单个处理元素最多只能是一个分区的成员。此功能允许基于动态数据结构的运行时状态进行重新配置,并允许对断开连接和重叠的数据结构进行分区。对数据结构的操作有望带来显著的加速;不需要将数据复制到新配置的分区。GPA系统由N个处理元件/RAM对和一个提供访问P个控制处理器或微控制器的互连网络组成。在目前的技术条件下,N和P分别为64K和16是可行的。
{"title":"The GPA machine: a generally partitionable MSIMD architecture","authors":"T. Bridges","doi":"10.1109/FMPC.1990.89460","DOIUrl":"https://doi.org/10.1109/FMPC.1990.89460","url":null,"abstract":"The GPA machine, a massively parallel, multiple single-instruction-stream-multiple-data-stream (MSIMD) system is described. Its distinguishing characteristics is the generality of its partitioning capabilities. Like the PASM system it can be dynamically reconfigured to operate as one or more independent SIMD machines. However, unlike PASM, the only constraint placed on partitioning is that an individual processing element is a member of at most one partition. This capability allows for reconfiguration based on the run-time status of dynamic data structures and for partitioning of disconnected and overlapping data structures. Significant speedups are expected from operating on data structures in place; copying of data to a newly configured partition is unnecessary. The GPA system consists of N processing-element/RAM pairs and an interconnection network providing access to and from P control processors or microcontrollers. With current technologies, values for N and P of 64K and 16, respectively, are feasible.<<ETX>>","PeriodicalId":193332,"journal":{"name":"[1990 Proceedings] The Third Symposium on the Frontiers of Massively Parallel Computation","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125532124","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
期刊
[1990 Proceedings] The Third Symposium on the Frontiers of Massively Parallel Computation
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1