{"title":"A parallel implementation of the chemically reacting CFD code, SPARK","authors":"J. C. Otto","doi":"10.1109/SHPCC.1992.232674","DOIUrl":null,"url":null,"abstract":"Describes a parallel version of the two-dimensional, chemically reacting CFD code, SPARK. The sequential code has been ported to run on the Intel iPSC/860-based parallel computers. Routines have been added to the code which partition the problem based on the global mesh, and then assign the resulting subdomains across the processors. Two subdomain mappings have been considered. The routines which compute spatial derivatives and the routine which adds artificial viscosity to the discretization were modified to handle the subdomain boundaries interior to the global domain, and an effort has been made to overlap the required communication/computation. Measurements of the performance of the code have been made for two test problems exercising all of the available options of the parallel code thus far. While the parallel efficiency of the code is quite good, the single-node performance has been much lower than expected for this architecture.<<ETX>>","PeriodicalId":254515,"journal":{"name":"Proceedings Scalable High Performance Computing Conference SHPCC-92.","volume":"29 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1992-04-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings Scalable High Performance Computing Conference SHPCC-92.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SHPCC.1992.232674","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Describes a parallel version of the two-dimensional, chemically reacting CFD code, SPARK. The sequential code has been ported to run on the Intel iPSC/860-based parallel computers. Routines have been added to the code which partition the problem based on the global mesh, and then assign the resulting subdomains across the processors. Two subdomain mappings have been considered. The routines which compute spatial derivatives and the routine which adds artificial viscosity to the discretization were modified to handle the subdomain boundaries interior to the global domain, and an effort has been made to overlap the required communication/computation. Measurements of the performance of the code have been made for two test problems exercising all of the available options of the parallel code thus far. While the parallel efficiency of the code is quite good, the single-node performance has been much lower than expected for this architecture.<>