首页 > 最新文献

Proceedings the Ninth International Symposium on High-Performance Distributed Computing最新文献

英文 中文
Bypass: a tool for building split execution systems 旁路:用于构建分割执行系统的工具
D. Thain, M. Livny
Split execution is a common model for providing a friendly environment on a foreign machine. In this model, a remotely executing process sends some or all of its system calls back to a home environment for execution. Unfortunately, hand-coding split execution systems for experimentation and research is difficult and error-prone. We have built a tool, called Bypass, for quickly producing portable and correct split execution systems for unmodified legacy applications. We demonstrate Bypass by using it to transparently connect a POSIX application to a simple data staging system based on the Globus toolkit.
分割执行是在外部机器上提供友好环境的常用模型。在此模型中,远程执行进程将其部分或全部系统调用发送回主环境执行。不幸的是,手工编码用于实验和研究的分割执行系统是困难的,而且容易出错。我们已经构建了一个名为Bypass的工具,用于为未修改的遗留应用程序快速生成可移植且正确的分割执行系统。我们通过使用它将POSIX应用程序透明地连接到基于Globus工具包的简单数据分级系统来演示Bypass。
{"title":"Bypass: a tool for building split execution systems","authors":"D. Thain, M. Livny","doi":"10.1109/HPDC.2000.868637","DOIUrl":"https://doi.org/10.1109/HPDC.2000.868637","url":null,"abstract":"Split execution is a common model for providing a friendly environment on a foreign machine. In this model, a remotely executing process sends some or all of its system calls back to a home environment for execution. Unfortunately, hand-coding split execution systems for experimentation and research is difficult and error-prone. We have built a tool, called Bypass, for quickly producing portable and correct split execution systems for unmodified legacy applications. We demonstrate Bypass by using it to transparently connect a POSIX application to a simple data staging system based on the Globus toolkit.","PeriodicalId":400728,"journal":{"name":"Proceedings the Ninth International Symposium on High-Performance Distributed Computing","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123570578","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
The Cactus Code: a problem solving environment for the grid Cactus Code:一个解决网格问题的环境
Gabrielle Allen, W. Benger, T. Goodale, H. Hege, Gerd Lanfermann, André Merzky, T. Radke, E. Seidel, J. Shalf
Cactus is an open source problem solving environment designed for scientists and engineers. Its modular structure facilitates parallel computation across different architectures and collaborative code development between different groups. The Cactus Code originated in the academic research community, where it has been developed and used over many years by a large international collaboration of physicists and computational scientists. We discuss how the intensive computing requirements of physics applications now using the Cactus Code encourage the use of distributed and metacomputing, describe the development and experiments which have already been performed with Cactus, and detail how its design makes it an ideal application test-bed for Grid computing.
Cactus是一个为科学家和工程师设计的开源问题解决环境。它的模块化结构促进了不同体系结构之间的并行计算和不同组之间的协作代码开发。Cactus Code起源于学术研究社区,由物理学家和计算科学家组成的大型国际合作组织开发和使用了多年。我们讨论了现在使用Cactus代码的物理应用程序的密集计算需求如何鼓励使用分布式和元计算,描述了Cactus已经执行的开发和实验,并详细说明了它的设计如何使其成为网格计算的理想应用程序测试平台。
{"title":"The Cactus Code: a problem solving environment for the grid","authors":"Gabrielle Allen, W. Benger, T. Goodale, H. Hege, Gerd Lanfermann, André Merzky, T. Radke, E. Seidel, J. Shalf","doi":"10.1109/HPDC.2000.868657","DOIUrl":"https://doi.org/10.1109/HPDC.2000.868657","url":null,"abstract":"Cactus is an open source problem solving environment designed for scientists and engineers. Its modular structure facilitates parallel computation across different architectures and collaborative code development between different groups. The Cactus Code originated in the academic research community, where it has been developed and used over many years by a large international collaboration of physicists and computational scientists. We discuss how the intensive computing requirements of physics applications now using the Cactus Code encourage the use of distributed and metacomputing, describe the development and experiments which have already been performed with Cactus, and detail how its design makes it an ideal application test-bed for Grid computing.","PeriodicalId":400728,"journal":{"name":"Proceedings the Ninth International Symposium on High-Performance Distributed Computing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130318478","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 135
Uintah: a massively parallel problem solving environment untah:一个大规模并行解决问题的环境
J., Davison de St. Germain, J. McCorquodale, S. Parker, Christopher R. Johnson
Describes Uintah, a component-based visual problem-solving environment (PSE) that is designed to specifically address the unique problems of massively parallel computation on tera-scale computing platforms. Uintah supports the entire life-cycle of scientific applications by allowing scientific programmers to quickly and easily develop new techniques, debug new implementations and apply known algorithms to solve novel problems. Uintah is built on three principles: (1) as much as possible, the complexities of parallel execution should be handled for the scientist, (2) the software should be reusable at the component level, and (3) scientists should be able to dynamically steer and visualize their simulation results as the simulation executes. To provide this functionality, Uintah builds upon the best features of the SCIRun (Scientific Computing and Imaging Run-time) PSE and the DoE (Department of Energy) Common Component Architecture (CCA).
描述了一个基于组件的可视化问题解决环境(PSE),专门用于解决在太大规模计算平台上大规模并行计算的独特问题。untah支持科学应用的整个生命周期,允许科学程序员快速轻松地开发新技术,调试新实现并应用已知算法来解决新问题。inttah建立在三个原则之上:(1)尽可能多地为科学家处理并行执行的复杂性,(2)软件应该在组件级别上可重用,(3)科学家应该能够在仿真执行时动态地引导和可视化他们的仿真结果。为了提供这一功能,inttah建立在SCIRun(科学计算和成像运行时)PSE和DoE(能源部)通用组件体系结构(CCA)的最佳特性之上。
{"title":"Uintah: a massively parallel problem solving environment","authors":"J., Davison de St. Germain, J. McCorquodale, S. Parker, Christopher R. Johnson","doi":"10.1109/HPDC.2000.868632","DOIUrl":"https://doi.org/10.1109/HPDC.2000.868632","url":null,"abstract":"Describes Uintah, a component-based visual problem-solving environment (PSE) that is designed to specifically address the unique problems of massively parallel computation on tera-scale computing platforms. Uintah supports the entire life-cycle of scientific applications by allowing scientific programmers to quickly and easily develop new techniques, debug new implementations and apply known algorithms to solve novel problems. Uintah is built on three principles: (1) as much as possible, the complexities of parallel execution should be handled for the scientist, (2) the software should be reusable at the component level, and (3) scientists should be able to dynamically steer and visualize their simulation results as the simulation executes. To provide this functionality, Uintah builds upon the best features of the SCIRun (Scientific Computing and Imaging Run-time) PSE and the DoE (Department of Energy) Common Component Architecture (CCA).","PeriodicalId":400728,"journal":{"name":"Proceedings the Ninth International Symposium on High-Performance Distributed Computing","volume":"111 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129818459","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 195
Grid-based file access: the Legion I/O model 基于网格的文件访问:军团I/O模型
B. White, A. Grimshaw, A. Nguyen-Tuong
The unprecedented scale, heterogeneity and varied usage patterns of computational grids pose significant technical challenges to any underlying file system that supports them. While grids present a host of new concerns for file access, we focus on two issues: performance and usability. We discuss the Legion I/O model and interface to address the latter area. We compare the Legion and Globus I/O models against a baseline to validate the efficiency of existent grid-based file access solutions.
计算网格前所未有的规模、异构性和不同的使用模式给支持它们的任何底层文件系统带来了重大的技术挑战。虽然网格为文件访问提出了许多新的问题,但我们主要关注两个问题:性能和可用性。我们将讨论军团I/O模型和接口来解决后一个问题。我们将Legion和Globus I/O模型与基线进行比较,以验证现有的基于网格的文件访问解决方案的效率。
{"title":"Grid-based file access: the Legion I/O model","authors":"B. White, A. Grimshaw, A. Nguyen-Tuong","doi":"10.1109/HPDC.2000.868647","DOIUrl":"https://doi.org/10.1109/HPDC.2000.868647","url":null,"abstract":"The unprecedented scale, heterogeneity and varied usage patterns of computational grids pose significant technical challenges to any underlying file system that supports them. While grids present a host of new concerns for file access, we focus on two issues: performance and usability. We discuss the Legion I/O model and interface to address the latter area. We compare the Legion and Globus I/O models against a baseline to validate the efficiency of existent grid-based file access solutions.","PeriodicalId":400728,"journal":{"name":"Proceedings the Ninth International Symposium on High-Performance Distributed Computing","volume":"70 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114280019","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 60
Using idle workstations to implement predictive prefetching 利用空闲工作站实现预测性预取
Jasmine Y. Q. Wang, J. Ong, Y. Coady, M. Feeley
The benefits of Markov-based predictive prefetching have been largely overshadowed by the overhead required to produce high-quality predictions. While both theoretical and simulation results for prediction algorithms appear promising, substantial limitations exist in practice. This outcome can be partially attributed to the fact that practical implementations ultimately make compromises in order to reduce overhead. These compromises limit the level of algorithm complexity, the variety of access patterns and the granularity of trace data that the implementation supports. This paper describes the design and implementation of GMS-3P (Global Memory System with Parallel Predictive Prefetching), an operating system kernel extension that offloads prediction overhead to idle network nodes. GMS-3P builds on the GMS global memory system, which pages to and from remote workstation memory. In GMS-3P, the target node sends an online trace of an application's page faults to an idle node that is running a Markov-based prediction algorithm. The prediction node then uses GMS to prefetch pages to the target node from the memory of other workstations in the network. Our preliminary results show that predictive prefetching can reduce the remote-memory page fault time by 60% or more and that, by offloading prediction overhead to an idle node, GMS-3P can reduce this improved latency by between 24% and 44%, depending on the Markov model order.
基于马尔可夫的预测预取的好处在很大程度上被产生高质量预测所需的开销所掩盖。虽然预测算法的理论和仿真结果都很有希望,但在实践中存在很大的局限性。这一结果可以部分归因于实际实现最终为了减少开销而做出妥协的事实。这些妥协限制了算法的复杂性、访问模式的多样性以及实现所支持的跟踪数据的粒度。本文描述了GMS-3P (Global Memory System with Parallel Predictive prefetch)的设计和实现,GMS-3P是一个操作系统内核扩展,可以将预测开销转移到空闲的网络节点上。GMS- 3p建立在GMS全局内存系统之上,该系统在远程工作站内存之间进行分页。在GMS-3P中,目标节点将应用程序页面错误的在线跟踪发送到正在运行基于markov的预测算法的空闲节点。然后,预测节点使用GMS从网络中其他工作站的内存中预取页面到目标节点。我们的初步结果表明,预测性预取可以将远程内存页面故障时间减少60%或更多,并且通过将预测开销卸载到空闲节点,GMS-3P可以将这种改进的延迟减少24%到44%,具体取决于马尔可夫模型的顺序。
{"title":"Using idle workstations to implement predictive prefetching","authors":"Jasmine Y. Q. Wang, J. Ong, Y. Coady, M. Feeley","doi":"10.1109/HPDC.2000.868638","DOIUrl":"https://doi.org/10.1109/HPDC.2000.868638","url":null,"abstract":"The benefits of Markov-based predictive prefetching have been largely overshadowed by the overhead required to produce high-quality predictions. While both theoretical and simulation results for prediction algorithms appear promising, substantial limitations exist in practice. This outcome can be partially attributed to the fact that practical implementations ultimately make compromises in order to reduce overhead. These compromises limit the level of algorithm complexity, the variety of access patterns and the granularity of trace data that the implementation supports. This paper describes the design and implementation of GMS-3P (Global Memory System with Parallel Predictive Prefetching), an operating system kernel extension that offloads prediction overhead to idle network nodes. GMS-3P builds on the GMS global memory system, which pages to and from remote workstation memory. In GMS-3P, the target node sends an online trace of an application's page faults to an idle node that is running a Markov-based prediction algorithm. The prediction node then uses GMS to prefetch pages to the target node from the memory of other workstations in the network. Our preliminary results show that predictive prefetching can reduce the remote-memory page fault time by 60% or more and that, by offloading prediction overhead to an idle node, GMS-3P can reduce this improved latency by between 24% and 44%, depending on the Markov model order.","PeriodicalId":400728,"journal":{"name":"Proceedings the Ninth International Symposium on High-Performance Distributed Computing","volume":"248 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117347983","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
CAVEStudy: an infrastructure for computational steering in virtual reality environments CAVEStudy:虚拟现实环境中计算导向的基础设施
L. Renambot, H. Bal, D. Germans, H. Spoelder
We present the CAVEStudy system that enables scientists to interactively steer a simulation from a virtual reality (VR) environment. No modification to the source code is necessary. CAVEStudy allows interactive and immersive analysis of a simulation running on a remote computer. Using a high-level description of the simulation, the system generates the communication layer (based on CAVERN-Soft) needed to control the execution and to gather data at runtime. We describe three case-studies implemented with CAVEStudy: soccer simulation, diode laser simulation and molecular dynamics.
我们提出了洞穴研究系统,使科学家能够交互式地从虚拟现实(VR)环境中引导模拟。不需要修改源代码。CAVEStudy允许对远程计算机上运行的模拟进行交互式和沉浸式分析。使用模拟的高级描述,系统生成所需的通信层(基于cave - soft),以控制执行并在运行时收集数据。我们描述了用CAVEStudy实现的三个案例研究:足球模拟、二极管激光模拟和分子动力学。
{"title":"CAVEStudy: an infrastructure for computational steering in virtual reality environments","authors":"L. Renambot, H. Bal, D. Germans, H. Spoelder","doi":"10.1109/HPDC.2000.868655","DOIUrl":"https://doi.org/10.1109/HPDC.2000.868655","url":null,"abstract":"We present the CAVEStudy system that enables scientists to interactively steer a simulation from a virtual reality (VR) environment. No modification to the source code is necessary. CAVEStudy allows interactive and immersive analysis of a simulation running on a remote computer. Using a high-level description of the simulation, the system generates the communication layer (based on CAVERN-Soft) needed to control the execution and to gather data at runtime. We describe three case-studies implemented with CAVEStudy: soccer simulation, diode laser simulation and molecular dynamics.","PeriodicalId":400728,"journal":{"name":"Proceedings the Ninth International Symposium on High-Performance Distributed Computing","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123970820","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Evaluating cluster-based network servers 评估基于集群的网络服务器
R. Bianchini, E. V. Carrera
Uses analytic modeling and simulation to evaluate network servers implemented on clusters of workstations. More specifically, we model the potential benefits of locality-conscious request distribution within the cluster and evaluate the performance of a cluster-based server called L2S (Locality and Load-balancing Server) which we designed in light of our experience with the model. Our most important modeling results show that locality-conscious distribution on a 16-node cluster can increase server throughput with respect to a locality-oblivious server by up to seven-fold, depending on the average size of the files requested and on the size of the server's working set. Our simulation results demonstrate that L2S achieves throughput that is within 22% of the full potential of locality-conscious distribution on 16 nodes, outperforming and significantly outscaling the best-known locality-conscious server. Based on our results and on the fact that the files serviced by network servers are becoming larger and more numerous, we conclude that our locality-conscious network server should prove very useful for its performance, scalability and availability.
使用分析建模和仿真来评估在工作站集群上实现的网络服务器。更具体地说,我们对集群中位置感知请求分发的潜在好处进行了建模,并评估了基于集群的服务器L2S (Locality and Load-balancing server)的性能,该服务器是我们根据使用该模型的经验设计的。我们最重要的建模结果表明,相对于位置无关的服务器,16节点集群上的位置敏感分布可以将服务器吞吐量提高多达7倍,具体取决于所请求文件的平均大小和服务器工作集的大小。我们的模拟结果表明,L2S在16个节点上实现的吞吐量在位置意识分布的全部潜力的22%以内,优于并显著超过了最著名的位置意识服务器。基于我们的结果以及网络服务器服务的文件变得越来越大和越来越多的事实,我们得出结论,我们的位置感知网络服务器应该证明其性能、可伸缩性和可用性非常有用。
{"title":"Evaluating cluster-based network servers","authors":"R. Bianchini, E. V. Carrera","doi":"10.1109/HPDC.2000.868635","DOIUrl":"https://doi.org/10.1109/HPDC.2000.868635","url":null,"abstract":"Uses analytic modeling and simulation to evaluate network servers implemented on clusters of workstations. More specifically, we model the potential benefits of locality-conscious request distribution within the cluster and evaluate the performance of a cluster-based server called L2S (Locality and Load-balancing Server) which we designed in light of our experience with the model. Our most important modeling results show that locality-conscious distribution on a 16-node cluster can increase server throughput with respect to a locality-oblivious server by up to seven-fold, depending on the average size of the files requested and on the size of the server's working set. Our simulation results demonstrate that L2S achieves throughput that is within 22% of the full potential of locality-conscious distribution on 16 nodes, outperforming and significantly outscaling the best-known locality-conscious server. Based on our results and on the fact that the files serviced by network servers are becoming larger and more numerous, we conclude that our locality-conscious network server should prove very useful for its performance, scalability and availability.","PeriodicalId":400728,"journal":{"name":"Proceedings the Ninth International Symposium on High-Performance Distributed Computing","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124735269","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Performance evaluation of a firewall-compliant Globus-based wide-area cluster system 兼容防火墙的基于globus的广域集群系统的性能评估
Yoshio Tanaka, Motonori Hirano, M. Sato, H. Nakada, S. Sekiguchi
Presents a performance evaluation of a wide-area cluster system based on a firewall-enabled Globus metacomputing toolkit. In order to establish communication links beyond the firewall, we have designed and implemented a resource manager called RMF (Resource Manager beyond the Firewall) and the Nexus Proxy, which relays TCP communication links beyond the firewall. In order to extend the Globus metacomputing toolkit to become firewall-enabled, we have built the Nexus Proxy into the Globus toolkit. We have built a firewall-enabled Globus-based wide-area cluster system in Japan and run some benchmarks on it. In this paper, we report various performance results, such as the communication bandwidth and latencies obtained, as well as application performance involving a tree search problem. In a wide-area environment, the communication latency through the Nexus Proxy is approximately six times larger when compared to that of direct communications. As the message size increases, however, the communication overhead caused by the Nexus Proxy can be negligible. We have developed a tree search problem using MPICH-G. We used a self-scheduling algorithm, which is considered to be suitable for a distributed heterogeneous metacomputing environment since it performs dynamic load balancing with low overhead. The performance results indicate that the communication overhead caused by the Nexus Proxy is not a severe problem in metacomputing environments.
介绍了基于启用防火墙的Globus元计算工具包的广域集群系统的性能评估。为了在防火墙之外建立通信链接,我们设计并实现了一个名为RMF(防火墙之外的资源管理器)的资源管理器和Nexus Proxy,它在防火墙之外中继TCP通信链接。为了扩展Globus元计算工具包,使其支持防火墙,我们在Globus工具包中构建了Nexus Proxy。我们在日本构建了一个启用防火墙的基于globus的广域集群系统,并在其上运行了一些基准测试。在本文中,我们报告了各种性能结果,例如获得的通信带宽和延迟,以及涉及树搜索问题的应用程序性能。在广域环境中,与直接通信相比,通过Nexus Proxy进行通信的延迟大约是直接通信的六倍。但是,随着消息大小的增加,由Nexus Proxy引起的通信开销可以忽略不计。我们使用MPICH-G开发了一个树搜索问题。我们使用了一种自调度算法,该算法被认为适合于分布式异构元计算环境,因为它以低开销执行动态负载平衡。性能结果表明,在元计算环境中,由Nexus Proxy引起的通信开销不是一个严重的问题。
{"title":"Performance evaluation of a firewall-compliant Globus-based wide-area cluster system","authors":"Yoshio Tanaka, Motonori Hirano, M. Sato, H. Nakada, S. Sekiguchi","doi":"10.1109/HPDC.2000.868642","DOIUrl":"https://doi.org/10.1109/HPDC.2000.868642","url":null,"abstract":"Presents a performance evaluation of a wide-area cluster system based on a firewall-enabled Globus metacomputing toolkit. In order to establish communication links beyond the firewall, we have designed and implemented a resource manager called RMF (Resource Manager beyond the Firewall) and the Nexus Proxy, which relays TCP communication links beyond the firewall. In order to extend the Globus metacomputing toolkit to become firewall-enabled, we have built the Nexus Proxy into the Globus toolkit. We have built a firewall-enabled Globus-based wide-area cluster system in Japan and run some benchmarks on it. In this paper, we report various performance results, such as the communication bandwidth and latencies obtained, as well as application performance involving a tree search problem. In a wide-area environment, the communication latency through the Nexus Proxy is approximately six times larger when compared to that of direct communications. As the message size increases, however, the communication overhead caused by the Nexus Proxy can be negligible. We have developed a tree search problem using MPICH-G. We used a self-scheduling algorithm, which is considered to be suitable for a distributed heterogeneous metacomputing environment since it performs dynamic load balancing with low overhead. The performance results indicate that the communication overhead caused by the Nexus Proxy is not a severe problem in metacomputing environments.","PeriodicalId":400728,"journal":{"name":"Proceedings the Ninth International Symposium on High-Performance Distributed Computing","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125197449","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Constructing the ASCI computational grid 构建ASCI计算网格
J. Beiriger, Wilbur R. Johnson, H. Bivens, S. Humphreys, R. Rhea
The Accelerated Strategic Computing Initiative (ASCI) computational grid is being constructed to interconnect the high performance computing resources of the nuclear weapons complex. The grid will simplify access to the diverse computing, storage, network, and visualization resources, and will enable the coordinated use of shared resources regardless of location. To match existing hardware platforms, required security services, and current simulation practices, the Globus MetaComputing Toolkit was selected to provide core grid services. The ASCI grid extends Globus functionality by operating as an independent grid, incorporating Kerberos-based security, interfacing to Sandia's Cplant/sup TM/, and extending job monitoring services. To fully meet ASCI's needs, the architecture layers distributed work management and criteria-driven resource selection services on top of Globus. These services simplify the grid interface by allowing users to simply request "run code X anywhere". This paper describes the initial design and prototype of the ASCI grid.
正在构建加速战略计算计划(ASCI)计算网格,以互连核武器综合体的高性能计算资源。网格将简化对各种计算、存储、网络和可视化资源的访问,并将使共享资源的协调使用成为可能,而不管位置如何。为了匹配现有的硬件平台、所需的安全服务和当前的模拟实践,选择Globus metaccomputing Toolkit来提供核心网格服务。ASCI网格扩展了Globus的功能,它作为一个独立的网格运行,结合了基于kerberos的安全性,与Sandia的plant/sup TM/接口,并扩展了作业监控服务。为了完全满足ASCI的需求,体系结构层在Globus之上分布了工作管理和标准驱动的资源选择服务。这些服务通过允许用户简单地请求“在任何地方运行代码X”来简化网格接口。本文介绍了ASCI网格的初步设计和原型。
{"title":"Constructing the ASCI computational grid","authors":"J. Beiriger, Wilbur R. Johnson, H. Bivens, S. Humphreys, R. Rhea","doi":"10.1109/HPDC.2000.868650","DOIUrl":"https://doi.org/10.1109/HPDC.2000.868650","url":null,"abstract":"The Accelerated Strategic Computing Initiative (ASCI) computational grid is being constructed to interconnect the high performance computing resources of the nuclear weapons complex. The grid will simplify access to the diverse computing, storage, network, and visualization resources, and will enable the coordinated use of shared resources regardless of location. To match existing hardware platforms, required security services, and current simulation practices, the Globus MetaComputing Toolkit was selected to provide core grid services. The ASCI grid extends Globus functionality by operating as an independent grid, incorporating Kerberos-based security, interfacing to Sandia's Cplant/sup TM/, and extending job monitoring services. To fully meet ASCI's needs, the architecture layers distributed work management and criteria-driven resource selection services on top of Globus. These services simplify the grid interface by allowing users to simply request \"run code X anywhere\". This paper describes the initial design and prototype of the ASCI grid.","PeriodicalId":400728,"journal":{"name":"Proceedings the Ninth International Symposium on High-Performance Distributed Computing","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126996112","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 42
A local area system network RHiNET-1: a network for high performance parallel computing 一个局域系统网络RHiNET-1:用于高性能并行计算的网络
H. Nishi, K. Tasho, J. Yamamoto, T. Kudoh, H. Amano
The Real World Computing Partnership (RWCP) has developed a local area system network (LASN) called RHiNET-1 (RWCP High-performance NETwork, version 1) using 1.33-Gbps optical interconnections for high-performance computing using personal computers distributed in an office or laboratory environment. The network interface, RHiNET-1/NI, uses a complex programmable logic device (CPLD) based protocol controller to provide an easy evaluation platform for various protocols. It fits in a 32-bit/33-MHz PCI bus. The switch, RHiNET-1/SW, consists of a single-chip CMOS switch and external SRAM. It provides low-latency, reliable communication with a flexible topology design. We are currently evaluating protocols on RHiNET-1. RHiNET-1 will enable a new form of high-performance computing environment. We are also developing the second implementation, RHiNET-2. RHiNET-2/NI will support a 64-bit/66-MHz PCI bus. RHiNET-2/SW is an 8-Gbps/port 8/spl times/8 single-chip ASIC switch. The aggregate bandwidth of RHiNET-2/SW is 64 Gbps.
真实世界计算伙伴关系(RWCP)已经开发了一种称为RHiNET-1 (RWCP高性能网络,版本1)的局域网系统网络(LASN),使用1.33 gbps光互连,用于使用分布在办公室或实验室环境中的个人计算机进行高性能计算。网络接口RHiNET-1/NI采用基于复杂可编程逻辑器件(CPLD)的协议控制器,为各种协议提供了一个简单的评估平台。它适合32位/33 mhz PCI总线。该开关,RHiNET-1/SW,由单片CMOS开关和外部SRAM组成。它通过灵活的拓扑设计提供低延迟、可靠的通信。我们目前正在评估RHiNET-1的协议。RHiNET-1将实现一种新型的高性能计算环境。我们还在开发第二个实现,RHiNET-2。RHiNET-2/NI将支持64位/66 mhz PCI总线。RHiNET-2/SW是一款8gbps /端口8/spl倍/8单芯片ASIC开关。RHiNET-2/SW的总带宽为64gbps。
{"title":"A local area system network RHiNET-1: a network for high performance parallel computing","authors":"H. Nishi, K. Tasho, J. Yamamoto, T. Kudoh, H. Amano","doi":"10.1109/HPDC.2000.868665","DOIUrl":"https://doi.org/10.1109/HPDC.2000.868665","url":null,"abstract":"The Real World Computing Partnership (RWCP) has developed a local area system network (LASN) called RHiNET-1 (RWCP High-performance NETwork, version 1) using 1.33-Gbps optical interconnections for high-performance computing using personal computers distributed in an office or laboratory environment. The network interface, RHiNET-1/NI, uses a complex programmable logic device (CPLD) based protocol controller to provide an easy evaluation platform for various protocols. It fits in a 32-bit/33-MHz PCI bus. The switch, RHiNET-1/SW, consists of a single-chip CMOS switch and external SRAM. It provides low-latency, reliable communication with a flexible topology design. We are currently evaluating protocols on RHiNET-1. RHiNET-1 will enable a new form of high-performance computing environment. We are also developing the second implementation, RHiNET-2. RHiNET-2/NI will support a 64-bit/66-MHz PCI bus. RHiNET-2/SW is an 8-Gbps/port 8/spl times/8 single-chip ASIC switch. The aggregate bandwidth of RHiNET-2/SW is 64 Gbps.","PeriodicalId":400728,"journal":{"name":"Proceedings the Ninth International Symposium on High-Performance Distributed Computing","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115933576","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
Proceedings the Ninth International Symposium on High-Performance Distributed Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1