S. L. Mirtaheri, Ehsan Mousavi Khaneghah, M. Sharifi, M. A. Azgomi
{"title":"高效消息传递机制对高性能分布式科学计算的影响","authors":"S. L. Mirtaheri, Ehsan Mousavi Khaneghah, M. Sharifi, M. A. Azgomi","doi":"10.1109/ISPA.2008.131","DOIUrl":null,"url":null,"abstract":"Parallel programming and distributed programming are two solutions for scientific applications to provide high performance and fast response time in parallel systems and distributed systems. Parallel and distributed systems must provide inter process communication (IPC) mechanisms like message passing mechanism as underlying platforms to enable communication between local and especially geographically dispersed and physically distributed processes. Communication overhead is the major problem in these systems and there are a lot of efforts to develop more efficient message passing mechanisms or to improve the network communication speed. This paper provides hard evidence that an efficient implementation of message passing mechanism on multi-computers reduces the execution time of a molecular dynamics code. A well-known program for macromolecular dynamics and mechanics called CHARMm is executed on a networked cluster. The performance of CHARMm is measured with two distributed implementations of message passing, namely a kernel-level implementation called DIPC2006 and a renowned library level implementation called MPI. It is shown that the performance of CHARMm on a DIPC2006 configured cluster is by far better than its performance on an optimized MPI configured similar cluster. Even ignoring the favorable points of kernel-level implementations, like safety, privilege, reliability, and primitiveness, the insight is twofold. Scientists are nowadays faced with more computational complexity and look for more efficient systems and mechanisms. Efficient distributed IPC mechanisms have direct effect on running scientistspsila simulations faster, and computer engineers may try harder to develop more efficient distributed implementations of IPC.","PeriodicalId":345341,"journal":{"name":"2008 IEEE International Symposium on Parallel and Distributed Processing with Applications","volume":"102 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2008-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"The Influence of Efficient Message Passing Mechanisms on High Performance Distributed Scientific Computing\",\"authors\":\"S. L. Mirtaheri, Ehsan Mousavi Khaneghah, M. Sharifi, M. A. Azgomi\",\"doi\":\"10.1109/ISPA.2008.131\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Parallel programming and distributed programming are two solutions for scientific applications to provide high performance and fast response time in parallel systems and distributed systems. Parallel and distributed systems must provide inter process communication (IPC) mechanisms like message passing mechanism as underlying platforms to enable communication between local and especially geographically dispersed and physically distributed processes. Communication overhead is the major problem in these systems and there are a lot of efforts to develop more efficient message passing mechanisms or to improve the network communication speed. This paper provides hard evidence that an efficient implementation of message passing mechanism on multi-computers reduces the execution time of a molecular dynamics code. A well-known program for macromolecular dynamics and mechanics called CHARMm is executed on a networked cluster. The performance of CHARMm is measured with two distributed implementations of message passing, namely a kernel-level implementation called DIPC2006 and a renowned library level implementation called MPI. It is shown that the performance of CHARMm on a DIPC2006 configured cluster is by far better than its performance on an optimized MPI configured similar cluster. Even ignoring the favorable points of kernel-level implementations, like safety, privilege, reliability, and primitiveness, the insight is twofold. Scientists are nowadays faced with more computational complexity and look for more efficient systems and mechanisms. Efficient distributed IPC mechanisms have direct effect on running scientistspsila simulations faster, and computer engineers may try harder to develop more efficient distributed implementations of IPC.\",\"PeriodicalId\":345341,\"journal\":{\"name\":\"2008 IEEE International Symposium on Parallel and Distributed Processing with Applications\",\"volume\":\"102 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2008-12-10\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2008 IEEE International Symposium on Parallel and Distributed Processing with Applications\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ISPA.2008.131\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2008 IEEE International Symposium on Parallel and Distributed Processing with Applications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISPA.2008.131","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
The Influence of Efficient Message Passing Mechanisms on High Performance Distributed Scientific Computing
Parallel programming and distributed programming are two solutions for scientific applications to provide high performance and fast response time in parallel systems and distributed systems. Parallel and distributed systems must provide inter process communication (IPC) mechanisms like message passing mechanism as underlying platforms to enable communication between local and especially geographically dispersed and physically distributed processes. Communication overhead is the major problem in these systems and there are a lot of efforts to develop more efficient message passing mechanisms or to improve the network communication speed. This paper provides hard evidence that an efficient implementation of message passing mechanism on multi-computers reduces the execution time of a molecular dynamics code. A well-known program for macromolecular dynamics and mechanics called CHARMm is executed on a networked cluster. The performance of CHARMm is measured with two distributed implementations of message passing, namely a kernel-level implementation called DIPC2006 and a renowned library level implementation called MPI. It is shown that the performance of CHARMm on a DIPC2006 configured cluster is by far better than its performance on an optimized MPI configured similar cluster. Even ignoring the favorable points of kernel-level implementations, like safety, privilege, reliability, and primitiveness, the insight is twofold. Scientists are nowadays faced with more computational complexity and look for more efficient systems and mechanisms. Efficient distributed IPC mechanisms have direct effect on running scientistspsila simulations faster, and computer engineers may try harder to develop more efficient distributed implementations of IPC.