J. Parker, T. Cwik, R. Ferraro, P. Liewer, P. Lyster, J. Patterson
{"title":"Helmholtz有限元在Mark III和Intel iPSC/860 Hypercubes上的性能","authors":"J. Parker, T. Cwik, R. Ferraro, P. Liewer, P. Lyster, J. Patterson","doi":"10.1109/DMCC.1991.633158","DOIUrl":null,"url":null,"abstract":"The large distributed memory capacities of hypercube computers are exploited by a finite element application which computes the scattered electromagetic field from heterogeneous objects with size large compared to a wavelength. Such problems scale well with hypercube dimension fo r large objects: by using the Recursive Inertial Partitioning algorithm and an iterative solver, the work done by each processor is nearly equal and communication overhead for the system set-up and solution is low. The application has been integrated into a user-friendly eirvironment on a graphics workstation in a local area network including hypercube host machines. Users need never know their solutions are obtained via a parallel computer. Scaling is shown by computing solutions for a series of models which double the number of variables for each increment of hypercube dimension. Timings are compared for the JPLICaltech Mark IIIfp Hypercube and the Intel iPSCI860 hypercube. Acceptable quality of solutions is obtained for object domains of hundreds of square wavelengths and resulting sparse matrix systems with order of 100,000 complex unknowns.","PeriodicalId":313314,"journal":{"name":"The Sixth Distributed Memory Computing Conference, 1991. Proceedings","volume":"16 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1991-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Helmholtz Finite Elements Performance On Mark III and Intel iPSC/860 Hypercubes\",\"authors\":\"J. Parker, T. Cwik, R. Ferraro, P. Liewer, P. Lyster, J. Patterson\",\"doi\":\"10.1109/DMCC.1991.633158\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The large distributed memory capacities of hypercube computers are exploited by a finite element application which computes the scattered electromagetic field from heterogeneous objects with size large compared to a wavelength. Such problems scale well with hypercube dimension fo r large objects: by using the Recursive Inertial Partitioning algorithm and an iterative solver, the work done by each processor is nearly equal and communication overhead for the system set-up and solution is low. The application has been integrated into a user-friendly eirvironment on a graphics workstation in a local area network including hypercube host machines. Users need never know their solutions are obtained via a parallel computer. Scaling is shown by computing solutions for a series of models which double the number of variables for each increment of hypercube dimension. Timings are compared for the JPLICaltech Mark IIIfp Hypercube and the Intel iPSCI860 hypercube. Acceptable quality of solutions is obtained for object domains of hundreds of square wavelengths and resulting sparse matrix systems with order of 100,000 complex unknowns.\",\"PeriodicalId\":313314,\"journal\":{\"name\":\"The Sixth Distributed Memory Computing Conference, 1991. Proceedings\",\"volume\":\"16 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"1991-04-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"The Sixth Distributed Memory Computing Conference, 1991. Proceedings\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/DMCC.1991.633158\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"The Sixth Distributed Memory Computing Conference, 1991. Proceedings","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/DMCC.1991.633158","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2
摘要
利用超立方体计算机庞大的分布式存储容量,实现了对尺寸大于波长的异质物体散射电磁场的有限元计算。这类问题在超立方体维度下可以很好地扩展到大型对象:通过使用递归惯性划分算法和迭代求解器,每个处理器所做的工作几乎相等,并且系统设置和解决方案的通信开销很低。该应用程序已集成到包括hypercube主机在内的局域网图形工作站的用户友好环境中。用户永远不需要知道他们的解是通过并行计算机得到的。通过计算一系列模型的解决方案来显示缩放,这些模型每增加一个超立方体维度,变量的数量就增加一倍。比较了JPLICaltech Mark IIIfp Hypercube和Intel iPSCI860 Hypercube的时序。对于数百平方波长的目标域和100,000阶复杂未知数的稀疏矩阵系统,获得了可接受的解质量。
Helmholtz Finite Elements Performance On Mark III and Intel iPSC/860 Hypercubes
The large distributed memory capacities of hypercube computers are exploited by a finite element application which computes the scattered electromagetic field from heterogeneous objects with size large compared to a wavelength. Such problems scale well with hypercube dimension fo r large objects: by using the Recursive Inertial Partitioning algorithm and an iterative solver, the work done by each processor is nearly equal and communication overhead for the system set-up and solution is low. The application has been integrated into a user-friendly eirvironment on a graphics workstation in a local area network including hypercube host machines. Users need never know their solutions are obtained via a parallel computer. Scaling is shown by computing solutions for a series of models which double the number of variables for each increment of hypercube dimension. Timings are compared for the JPLICaltech Mark IIIfp Hypercube and the Intel iPSCI860 hypercube. Acceptable quality of solutions is obtained for object domains of hundreds of square wavelengths and resulting sparse matrix systems with order of 100,000 complex unknowns.