首页 > 最新文献

International Journal of High Performance Computing Applications最新文献

英文 中文
Matrix-free approaches for GPU acceleration of a high-order finite element hydrodynamics application using MFEM, Umpire, and RAJA 使用MFEM、Umpire和RAJA的高阶有限元流体动力学应用的GPU加速的无矩阵方法
IF 3.1 3区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2021-12-14 DOI: 10.1177/10943420221100262
A. Vargas, T. Stitt, K. Weiss, V. Tomov, Jean-Sylvain Camier, T. Kolev, R. Rieben
With the introduction of advanced heterogeneous computing architectures based on GPU accelerators, large-scale production codes have had to rethink their numerical algorithms and incorporate new programming models and memory management strategies in order to run efficiently on the latest supercomputers. In this work we discuss our co-design strategy to address these challenges and achieve performance and portability with MARBL, a next-generation multi-physics code in development at Lawrence Livermore National Laboratory. We present a two-fold approach, wherein new hardware is used to motivate both new algorithms and new abstraction layers, resulting in a single source application code suitable for a variety of platforms. Focusing on MARBL’s ALE hydrodynamics package, we demonstrate scalability on different platforms and highlight that many of our innovations have been contributed back to open-source software libraries, such as MFEM (finite element algorithms) and RAJA (kernel abstractions).
随着基于GPU加速器的先进异构计算架构的引入,为了在最新的超级计算机上高效运行,大规模生产代码不得不重新考虑它们的数值算法,并结合新的编程模型和内存管理策略。在这项工作中,我们讨论了我们的共同设计策略,以解决这些挑战,并实现MARBL的性能和可移植性,MARBL是劳伦斯利弗莫尔国家实验室正在开发的下一代多物理场代码。我们提出了一种双重方法,其中使用新的硬件来激发新的算法和新的抽象层,从而产生适合各种平台的单一源代码应用程序代码。我们重点介绍了MARBL的ALE流体动力学包,展示了在不同平台上的可扩展性,并强调了我们的许多创新都贡献给了开源软件库,如MFEM(有限元算法)和RAJA(内核抽象)。
{"title":"Matrix-free approaches for GPU acceleration of a high-order finite element hydrodynamics application using MFEM, Umpire, and RAJA","authors":"A. Vargas, T. Stitt, K. Weiss, V. Tomov, Jean-Sylvain Camier, T. Kolev, R. Rieben","doi":"10.1177/10943420221100262","DOIUrl":"https://doi.org/10.1177/10943420221100262","url":null,"abstract":"With the introduction of advanced heterogeneous computing architectures based on GPU accelerators, large-scale production codes have had to rethink their numerical algorithms and incorporate new programming models and memory management strategies in order to run efficiently on the latest supercomputers. In this work we discuss our co-design strategy to address these challenges and achieve performance and portability with MARBL, a next-generation multi-physics code in development at Lawrence Livermore National Laboratory. We present a two-fold approach, wherein new hardware is used to motivate both new algorithms and new abstraction layers, resulting in a single source application code suitable for a variety of platforms. Focusing on MARBL’s ALE hydrodynamics package, we demonstrate scalability on different platforms and highlight that many of our innovations have been contributed back to open-source software libraries, such as MFEM (finite element algorithms) and RAJA (kernel abstractions).","PeriodicalId":54957,"journal":{"name":"International Journal of High Performance Computing Applications","volume":"36 1","pages":"492 - 509"},"PeriodicalIF":3.1,"publicationDate":"2021-12-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42558031","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Language models for the prediction of SARS-CoV-2 inhibitors 预测SARS-CoV-2抑制剂的语言模型
IF 3.1 3区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2021-12-14 DOI: 10.1101/2021.12.10.471928
Andrew E. Blanchard, John P. Gounley, D. Bhowmik, Mayanka Chandra Shekar, Isaac Lyngaas, Shang Gao, Junqi Yin, A. Tsaris, Feiyi Wang, J. Glaser
The COVID-19 pandemic highlights the need for computational tools to automate and accelerate drug design for novel protein targets. We leverage deep learning language models to generate and score drug candidates based on predicted protein binding affinity. We pre-trained a deep learning language model (BERT) on ∼9.6 billion molecules and achieved peak performance of 603 petaflops in mixed precision. Our work reduces pre-training time from days to hours, compared to previous efforts with this architecture, while also increasing the dataset size by nearly an order of magnitude. For scoring, we fine-tuned the language model using an assembled set of thousands of protein targets with binding affinity data and searched for inhibitors of specific protein targets, SARS-CoV-2 Mpro and PLpro. We utilized a genetic algorithm approach for finding optimal candidates using the generation and scoring capabilities of the language model. Our generalizable models accelerate the identification of inhibitors for emerging therapeutic targets.
COVID-19大流行凸显了对计算工具的需求,以自动化和加速针对新型蛋白质靶点的药物设计。我们利用深度学习语言模型根据预测的蛋白质结合亲和力生成和评分候选药物。我们在约96亿个分子上预训练了一个深度学习语言模型(BERT),并在混合精度下实现了603 petaflops的峰值性能。与之前使用这种架构的工作相比,我们的工作将预训练时间从几天减少到几个小时,同时还将数据集大小增加了近一个数量级。为了评分,我们使用数千个具有结合亲和力数据的蛋白质靶点集合对语言模型进行了精细调整,并寻找特定蛋白质靶点SARS-CoV-2 Mpro和PLpro的抑制剂。我们利用遗传算法的方法来寻找使用语言模型的生成和评分能力的最佳候选人。我们可推广的模型加速了新兴治疗靶点抑制剂的识别。
{"title":"Language models for the prediction of SARS-CoV-2 inhibitors","authors":"Andrew E. Blanchard, John P. Gounley, D. Bhowmik, Mayanka Chandra Shekar, Isaac Lyngaas, Shang Gao, Junqi Yin, A. Tsaris, Feiyi Wang, J. Glaser","doi":"10.1101/2021.12.10.471928","DOIUrl":"https://doi.org/10.1101/2021.12.10.471928","url":null,"abstract":"The COVID-19 pandemic highlights the need for computational tools to automate and accelerate drug design for novel protein targets. We leverage deep learning language models to generate and score drug candidates based on predicted protein binding affinity. We pre-trained a deep learning language model (BERT) on ∼9.6 billion molecules and achieved peak performance of 603 petaflops in mixed precision. Our work reduces pre-training time from days to hours, compared to previous efforts with this architecture, while also increasing the dataset size by nearly an order of magnitude. For scoring, we fine-tuned the language model using an assembled set of thousands of protein targets with binding affinity data and searched for inhibitors of specific protein targets, SARS-CoV-2 Mpro and PLpro. We utilized a genetic algorithm approach for finding optimal candidates using the generation and scoring capabilities of the language model. Our generalizable models accelerate the identification of inhibitors for emerging therapeutic targets.","PeriodicalId":54957,"journal":{"name":"International Journal of High Performance Computing Applications","volume":"51 1","pages":"587 - 602"},"PeriodicalIF":3.1,"publicationDate":"2021-12-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"62337418","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Development of a hardware-accelerated simulation kernel for ultra-high vacuum with Nvidia RTX GPUs 用Nvidia RTX GPU开发超高真空度硬件加速仿真内核
IF 3.1 3区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2021-12-11 DOI: 10.1177/10943420211056654
Pascal R Bähr, B. Lang, P. Ueberholz, M. Ady, R. Kersevan
Molflow+ is a Monte Carlo (MC) simulation software for ultra-high vacuum, mainly used to simulate pressure in particle accelerators. In this article, we present and discuss the design choices arising in a new implementation of its ray-tracing–based simulation unit for Nvidia RTX Graphics Processing Units (GPUs). The GPU simulation kernel was designed with Nvidia’s OptiX 7 API to make use of modern hardware-accelerated ray-tracing units, found in recent RTX series GPUs based on the Turing and Ampere architectures. Even with the challenges posed by switching to 32 bit computations, our kernel runs much faster than on comparable CPUs at the expense of a marginal drop in calculation precision.
Molflow+是一款用于超高真空的蒙特卡罗(MC)模拟软件,主要用于模拟粒子加速器中的压力。在本文中,我们介绍并讨论了Nvidia RTX图形处理单元(GPU)基于光线跟踪的模拟单元的新实现中出现的设计选择。GPU模拟内核是使用Nvidia的OptiX 7 API设计的,以利用现代硬件加速光线跟踪单元,这些单元在最近基于图灵和安培架构的RTX系列GPU中发现。即使切换到32位计算带来了挑战,我们的内核运行速度也比同等CPU快得多,代价是计算精度略有下降。
{"title":"Development of a hardware-accelerated simulation kernel for ultra-high vacuum with Nvidia RTX GPUs","authors":"Pascal R Bähr, B. Lang, P. Ueberholz, M. Ady, R. Kersevan","doi":"10.1177/10943420211056654","DOIUrl":"https://doi.org/10.1177/10943420211056654","url":null,"abstract":"Molflow+ is a Monte Carlo (MC) simulation software for ultra-high vacuum, mainly used to simulate pressure in particle accelerators. In this article, we present and discuss the design choices arising in a new implementation of its ray-tracing–based simulation unit for Nvidia RTX Graphics Processing Units (GPUs). The GPU simulation kernel was designed with Nvidia’s OptiX 7 API to make use of modern hardware-accelerated ray-tracing units, found in recent RTX series GPUs based on the Turing and Ampere architectures. Even with the challenges posed by switching to 32 bit computations, our kernel runs much faster than on comparable CPUs at the expense of a marginal drop in calculation precision.","PeriodicalId":54957,"journal":{"name":"International Journal of High Performance Computing Applications","volume":"36 1","pages":"141 - 152"},"PeriodicalIF":3.1,"publicationDate":"2021-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42921743","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Co-design in the Exascale Computing Project Exascale计算项目中的协同设计
IF 3.1 3区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2021-11-01 DOI: 10.1177/10943420211059380
T. Germann
We provide an overview of the six co-design centers within the U.S. Department of Energy’s Exascale Computing Project, each of which is described in more detail in a separate paper in this special issue. We also give a perspective on the evolution of computational co-design.
我们概述了美国能源部百亿亿级计算项目中的六个协同设计中心,在本期特刊的另一篇论文中对每个中心进行了更详细的描述。我们也给出了计算协同设计的演变的观点。
{"title":"Co-design in the Exascale Computing Project","authors":"T. Germann","doi":"10.1177/10943420211059380","DOIUrl":"https://doi.org/10.1177/10943420211059380","url":null,"abstract":"We provide an overview of the six co-design centers within the U.S. Department of Energy’s Exascale Computing Project, each of which is described in more detail in a separate paper in this special issue. We also give a perspective on the evolution of computational co-design.","PeriodicalId":54957,"journal":{"name":"International Journal of High Performance Computing Applications","volume":"35 1","pages":"503 - 507"},"PeriodicalIF":3.1,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43633066","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Semi-Lagrangian 4d, 5d, and 6d kinetic plasma simulation on large-scale GPU-equipped supercomputers 半拉格朗日4d, 5d和6d动力学等离子体在大型gpu超级计算机上的模拟
IF 3.1 3区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2021-10-27 DOI: 10.1177/10943420221137599
L. Einkemmer, A. Moriggl
Running kinetic plasma physics simulations using grid-based solvers is very demanding both in terms of memory as well as computational cost. This is primarily due to the up to six-dimensional phase space and the associated unfavorable scaling of the computational cost as a function of grid spacing (often termed the curse of dimensionality). In this article, we present 4d, 5d, and 6d simulations of the Vlasov–Poisson equation with a split-step semi-Lagrangian discontinuous Galerkin scheme on graphic processing units (GPUs). The local communication pattern of this method allows an efficient implementation on large-scale GPU-based systems and emphasizes the importance of considering algorithmic and high-performance computing aspects in unison. We demonstrate a single node performance above 2 TB/s effective memory bandwidth (on a node with four A100 GPUs) and show excellent scaling (parallel efficiency between 30% and 67%) for up to 1536 A100 GPUs on JUWELS Booster. Graphical Abstract
使用基于网格的求解器运行动态等离子体物理模拟在内存和计算成本方面都是非常苛刻的。这主要是由于高达六维的相位空间以及作为网格间距函数的计算成本的相关不利缩放(通常称为维度诅咒)。在本文中,我们在图形处理单元(gpu)上用分步半拉格朗日不连续伽辽金格式给出了Vlasov-Poisson方程的4d、5d和6d模拟。该方法的本地通信模式允许在基于gpu的大规模系统上有效实现,并强调了同时考虑算法和高性能计算方面的重要性。我们在JUWELS Booster上展示了超过2 TB/s有效内存带宽的单节点性能(在具有四个A100 gpu的节点上),并显示了高达1536个A100 gpu的出色扩展(并行效率在30%到67%之间)。图形抽象
{"title":"Semi-Lagrangian 4d, 5d, and 6d kinetic plasma simulation on large-scale GPU-equipped supercomputers","authors":"L. Einkemmer, A. Moriggl","doi":"10.1177/10943420221137599","DOIUrl":"https://doi.org/10.1177/10943420221137599","url":null,"abstract":"Running kinetic plasma physics simulations using grid-based solvers is very demanding both in terms of memory as well as computational cost. This is primarily due to the up to six-dimensional phase space and the associated unfavorable scaling of the computational cost as a function of grid spacing (often termed the curse of dimensionality). In this article, we present 4d, 5d, and 6d simulations of the Vlasov–Poisson equation with a split-step semi-Lagrangian discontinuous Galerkin scheme on graphic processing units (GPUs). The local communication pattern of this method allows an efficient implementation on large-scale GPU-based systems and emphasizes the importance of considering algorithmic and high-performance computing aspects in unison. We demonstrate a single node performance above 2 TB/s effective memory bandwidth (on a node with four A100 GPUs) and show excellent scaling (parallel efficiency between 30% and 67%) for up to 1536 A100 GPUs on JUWELS Booster. Graphical Abstract","PeriodicalId":54957,"journal":{"name":"International Journal of High Performance Computing Applications","volume":"37 1","pages":"180 - 196"},"PeriodicalIF":3.1,"publicationDate":"2021-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48720975","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
SAM++: Porting the E3SM-MMF cloud resolving model using a C++ portability library sam++:使用c++可移植性库移植E3SM-MMF云解析模型
IF 3.1 3区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2021-10-10 DOI: 10.1177/10943420211044495
Isaac Lyngaas, M. Norman, Youngsung Kim
In this work, we demonstrate the process for porting the cloud resolving model (CRM) used in the Energy Exascale Earth System Model Multi-Scale Modeling Framework (E3SM-MMF) from its original Fortran code base to C++ code using a portability library. This porting process is performed using the Yet Another Kernel Library (YAKL), a simplified C++ portability library that specializes in Fortran porting. In particular, we detail our step-by-step approach for porting the System for Atmospheric Modeling (SAM), the CRM used in E3SM-MMF, using a hybrid Fortran/C++ framework that allows for systematic reproduction and correctness testing of gradually ported YAKL C++ code. Additionally, analysis is done on the performance of the ported code using OLCF’s Summit supercomputer.
在这项工作中,我们演示了使用可移植性库将能源Exascale地球系统模型多尺度建模框架(E3SM-MMF)中使用的云解析模型(CRM)从其原始Fortran代码库移植到C++代码的过程。这个移植过程是使用另一个内核库(YAKL)执行的,YAKL是一个简化的C++可移植性库,专门用于Fortran移植。特别是,我们详细介绍了我们逐步移植大气建模系统(SAM)的方法,该系统是E3SM-MMF中使用的CRM,使用了一个混合的Fortran/C++框架,该框架允许系统复制和正确性测试逐步移植的YAKL C++代码。此外,使用OLCF的Summit超级计算机对移植代码的性能进行了分析。
{"title":"SAM++: Porting the E3SM-MMF cloud resolving model using a C++ portability library","authors":"Isaac Lyngaas, M. Norman, Youngsung Kim","doi":"10.1177/10943420211044495","DOIUrl":"https://doi.org/10.1177/10943420211044495","url":null,"abstract":"In this work, we demonstrate the process for porting the cloud resolving model (CRM) used in the Energy Exascale Earth System Model Multi-Scale Modeling Framework (E3SM-MMF) from its original Fortran code base to C++ code using a portability library. This porting process is performed using the Yet Another Kernel Library (YAKL), a simplified C++ portability library that specializes in Fortran porting. In particular, we detail our step-by-step approach for porting the System for Atmospheric Modeling (SAM), the CRM used in E3SM-MMF, using a hybrid Fortran/C++ framework that allows for systematic reproduction and correctness testing of gradually ported YAKL C++ code. Additionally, analysis is done on the performance of the ported code using OLCF’s Summit supercomputer.","PeriodicalId":54957,"journal":{"name":"International Journal of High Performance Computing Applications","volume":"36 1","pages":"214 - 230"},"PeriodicalIF":3.1,"publicationDate":"2021-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49062634","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
EXAGRAPH: Graph and combinatorial methods for enabling exascale applications EXAGRAPH:启用exascale应用程序的图形和组合方法
IF 3.1 3区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2021-09-30 DOI: 10.1177/10943420211029299
Seher Acer, A. Azad, E. Boman, A. Buluç, K. Devine, SM Ferdous, Nitin Gawande, Sayan Ghosh, M. Halappanavar, A. Kalyanaraman, Arif M. Khan, Marco Minutoli, A. Pothen, S. Rajamanickam, Oguz Selvitopi, Nathan R. Tallent, Antonino Tumeo
Combinatorial algorithms in general and graph algorithms in particular play a critical enabling role in numerous scientific applications. However, the irregular memory access nature of these algorithms makes them one of the hardest algorithmic kernels to implement on parallel systems. With tens of billions of hardware threads and deep memory hierarchies, the exascale computing systems in particular pose extreme challenges in scaling graph algorithms. The codesign center on combinatorial algorithms, ExaGraph, was established to design and develop methods and techniques for efficient implementation of key combinatorial (graph) algorithms chosen from a diverse set of exascale applications. Algebraic and combinatorial methods have a complementary role in the advancement of computational science and engineering, including playing an enabling role on each other. In this paper, we survey the algorithmic and software development activities performed under the auspices of ExaGraph from both a combinatorial and an algebraic perspective. In particular, we detail our recent efforts in porting the algorithms to manycore accelerator (GPU) architectures. We also provide a brief survey of the applications that have benefited from the scalable implementations of different combinatorial algorithms to enable scientific discovery at scale. We believe that several applications will benefit from the algorithmic and software tools developed by the ExaGraph team.
一般的组合算法,特别是图算法在许多科学应用中发挥着关键的作用。然而,这些算法的不规则内存访问特性使它们成为最难在并行系统上实现的算法内核之一。凭借数百亿的硬件线程和深度内存层次结构,exascale计算系统尤其对缩放图算法提出了极大的挑战。组合算法的代码设计中心ExaGraph的成立是为了设计和开发有效实现从一组不同的exascale应用程序中选择的关键组合(图)算法的方法和技术。代数方法和组合方法在计算科学和工程的发展中具有互补作用,包括相互发挥促进作用。在本文中,我们从组合和代数的角度考察了ExaGraph主持下进行的算法和软件开发活动。特别是,我们详细介绍了我们最近在将算法移植到多核加速器(GPU)架构方面所做的努力。我们还简要介绍了受益于不同组合算法的可扩展实现的应用程序,以实现大规模的科学发现。我们相信,ExaGraph团队开发的算法和软件工具将使一些应用程序受益。
{"title":"EXAGRAPH: Graph and combinatorial methods for enabling exascale applications","authors":"Seher Acer, A. Azad, E. Boman, A. Buluç, K. Devine, SM Ferdous, Nitin Gawande, Sayan Ghosh, M. Halappanavar, A. Kalyanaraman, Arif M. Khan, Marco Minutoli, A. Pothen, S. Rajamanickam, Oguz Selvitopi, Nathan R. Tallent, Antonino Tumeo","doi":"10.1177/10943420211029299","DOIUrl":"https://doi.org/10.1177/10943420211029299","url":null,"abstract":"Combinatorial algorithms in general and graph algorithms in particular play a critical enabling role in numerous scientific applications. However, the irregular memory access nature of these algorithms makes them one of the hardest algorithmic kernels to implement on parallel systems. With tens of billions of hardware threads and deep memory hierarchies, the exascale computing systems in particular pose extreme challenges in scaling graph algorithms. The codesign center on combinatorial algorithms, ExaGraph, was established to design and develop methods and techniques for efficient implementation of key combinatorial (graph) algorithms chosen from a diverse set of exascale applications. Algebraic and combinatorial methods have a complementary role in the advancement of computational science and engineering, including playing an enabling role on each other. In this paper, we survey the algorithmic and software development activities performed under the auspices of ExaGraph from both a combinatorial and an algebraic perspective. In particular, we detail our recent efforts in porting the algorithms to manycore accelerator (GPU) architectures. We also provide a brief survey of the applications that have benefited from the scalable implementations of different combinatorial algorithms to enable scientific discovery at scale. We believe that several applications will benefit from the algorithmic and software tools developed by the ExaGraph team.","PeriodicalId":54957,"journal":{"name":"International Journal of High Performance Computing Applications","volume":"35 1","pages":"553 - 571"},"PeriodicalIF":3.1,"publicationDate":"2021-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42735882","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Co-design Center for Exascale Machine Learning Technologies (ExaLearn) Exascale机器学习技术协同设计中心(ExaLearn)
IF 3.1 3区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2021-09-27 DOI: 10.1177/10943420211029302
Francis J. Alexander, James Ang, Jenna A. Bilbrey, J. Balewski, T. Casey, Ryan Chard, J. Choi, Sutanay Choudhury, B. Debusschere, Anthony Degennaro, Nikoli Dryden, J. Ellis, Ian T. Foster, Cristina Garcia Cardona, Sayan Ghosh, P. Harrington, Yunzhi Huang, S. Jha, Travis Johnston, Ai Kagawa, R. Kannan, Neeraj Kumar, Zhengchun Liu, N. Maruyama, S. Matsuoka, Erin McCarthy, J. Mohd-Yusof, Peter Nugent, Yosuke Oyama, T. Proffen, D. Pugmire, S. Rajamanickam, V. Ramakrishniah, M. Schram, S. Seal, G. Sivaraman, Christine M. Sweeney, Li Tan, R. Thakur, B. V. Van Essen, Logan T. Ward, P. Welch, Michael Wolf, S. Xantheas, K. Yager, Shinjae Yoo, Byung-Jun Yoon
Rapid growth in data, computational methods, and computing power is driving a remarkable revolution in what variously is termed machine learning (ML), statistical learning, computational learning, and artificial intelligence. In addition to highly visible successes in machine-based natural language translation, playing the game Go, and self-driving cars, these new technologies also have profound implications for computational and experimental science and engineering, as well as for the exascale computing systems that the Department of Energy (DOE) is developing to support those disciplines. Not only do these learning technologies open up exciting opportunities for scientific discovery on exascale systems, they also appear poised to have important implications for the design and use of exascale computers themselves, including high-performance computing (HPC) for ML and ML for HPC. The overarching goal of the ExaLearn co-design project is to provide exascale ML software for use by Exascale Computing Project (ECP) applications, other ECP co-design centers, and DOE experimental facilities and leadership class computing facilities.
数据、计算方法和计算能力的快速增长正在推动一场引人注目的革命,这场革命被称为机器学习(ML)、统计学习、计算学习和人工智能。除了在基于机器的自然语言翻译、围棋游戏和自动驾驶汽车方面取得了引人注目的成功外,这些新技术还对计算和实验科学与工程,以及能源部(DOE)为支持这些学科而开发的EB级计算系统产生了深远的影响。这些学习技术不仅为exascale系统的科学发现开辟了令人兴奋的机会,而且似乎也将对exascale计算机本身的设计和使用产生重要影响,包括用于ML的高性能计算(HPC)和用于HPC的ML。ExaLearn联合设计项目的首要目标是提供exascale ML软件,供exascale计算项目(ECP)应用程序、其他ECP联合设计中心、DOE实验设施和领导级计算设施使用。
{"title":"Co-design Center for Exascale Machine Learning Technologies (ExaLearn)","authors":"Francis J. Alexander, James Ang, Jenna A. Bilbrey, J. Balewski, T. Casey, Ryan Chard, J. Choi, Sutanay Choudhury, B. Debusschere, Anthony Degennaro, Nikoli Dryden, J. Ellis, Ian T. Foster, Cristina Garcia Cardona, Sayan Ghosh, P. Harrington, Yunzhi Huang, S. Jha, Travis Johnston, Ai Kagawa, R. Kannan, Neeraj Kumar, Zhengchun Liu, N. Maruyama, S. Matsuoka, Erin McCarthy, J. Mohd-Yusof, Peter Nugent, Yosuke Oyama, T. Proffen, D. Pugmire, S. Rajamanickam, V. Ramakrishniah, M. Schram, S. Seal, G. Sivaraman, Christine M. Sweeney, Li Tan, R. Thakur, B. V. Van Essen, Logan T. Ward, P. Welch, Michael Wolf, S. Xantheas, K. Yager, Shinjae Yoo, Byung-Jun Yoon","doi":"10.1177/10943420211029302","DOIUrl":"https://doi.org/10.1177/10943420211029302","url":null,"abstract":"Rapid growth in data, computational methods, and computing power is driving a remarkable revolution in what variously is termed machine learning (ML), statistical learning, computational learning, and artificial intelligence. In addition to highly visible successes in machine-based natural language translation, playing the game Go, and self-driving cars, these new technologies also have profound implications for computational and experimental science and engineering, as well as for the exascale computing systems that the Department of Energy (DOE) is developing to support those disciplines. Not only do these learning technologies open up exciting opportunities for scientific discovery on exascale systems, they also appear poised to have important implications for the design and use of exascale computers themselves, including high-performance computing (HPC) for ML and ML for HPC. The overarching goal of the ExaLearn co-design project is to provide exascale ML software for use by Exascale Computing Project (ECP) applications, other ECP co-design centers, and DOE experimental facilities and leadership class computing facilities.","PeriodicalId":54957,"journal":{"name":"International Journal of High Performance Computing Applications","volume":"35 1","pages":"598 - 616"},"PeriodicalIF":3.1,"publicationDate":"2021-09-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45120001","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
AI-driven multiscale simulations illuminate mechanisms of SARS-CoV-2 spike dynamics. 人工智能驱动的多尺度模拟揭示了 SARS-CoV-2 穗状动态的机制。
IF 3.1 3区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2021-09-01 DOI: 10.1177/10943420211006452
Lorenzo Casalino, Abigail C Dommer, Zied Gaieb, Emilia P Barros, Terra Sztain, Surl-Hee Ahn, Anda Trifan, Alexander Brace, Anthony T Bogetti, Austin Clyde, Heng Ma, Hyungro Lee, Matteo Turilli, Syma Khalid, Lillian T Chong, Carlos Simmerling, David J Hardy, Julio Dc Maia, James C Phillips, Thorsten Kurth, Abraham C Stern, Lei Huang, John D McCalpin, Mahidhar Tatineni, Tom Gibbs, John E Stone, Shantenu Jha, Arvind Ramanathan, Rommie E Amaro

We develop a generalizable AI-driven workflow that leverages heterogeneous HPC resources to explore the time-dependent dynamics of molecular systems. We use this workflow to investigate the mechanisms of infectivity of the SARS-CoV-2 spike protein, the main viral infection machinery. Our workflow enables more efficient investigation of spike dynamics in a variety of complex environments, including within a complete SARS-CoV-2 viral envelope simulation, which contains 305 million atoms and shows strong scaling on ORNL Summit using NAMD. We present several novel scientific discoveries, including the elucidation of the spike's full glycan shield, the role of spike glycans in modulating the infectivity of the virus, and the characterization of the flexible interactions between the spike and the human ACE2 receptor. We also demonstrate how AI can accelerate conformational sampling across different systems and pave the way for the future application of such methods to additional studies in SARS-CoV-2 and other molecular systems.

我们开发了一种可通用的人工智能驱动工作流,利用异构高性能计算资源来探索分子系统随时间变化的动力学。我们利用该工作流研究了 SARS-CoV-2 穗状病毒的感染机制,它是主要的病毒感染机制。我们的工作流能够在各种复杂环境中更有效地研究尖峰动态,包括在完整的 SARS-CoV-2 病毒包膜模拟中,该模拟包含 3.05 亿个原子,并在使用 NAMD 的 ORNL Summit 上显示出很强的扩展性。我们介绍了几项新的科学发现,包括阐明尖峰的完整聚糖屏蔽、尖峰聚糖在调节病毒传染性中的作用以及尖峰与人类 ACE2 受体之间灵活相互作用的特征。我们还展示了人工智能如何加速不同系统的构象取样,并为将来将这种方法应用于 SARS-CoV-2 和其他分子系统的其他研究铺平了道路。
{"title":"AI-driven multiscale simulations illuminate mechanisms of SARS-CoV-2 spike dynamics.","authors":"Lorenzo Casalino, Abigail C Dommer, Zied Gaieb, Emilia P Barros, Terra Sztain, Surl-Hee Ahn, Anda Trifan, Alexander Brace, Anthony T Bogetti, Austin Clyde, Heng Ma, Hyungro Lee, Matteo Turilli, Syma Khalid, Lillian T Chong, Carlos Simmerling, David J Hardy, Julio Dc Maia, James C Phillips, Thorsten Kurth, Abraham C Stern, Lei Huang, John D McCalpin, Mahidhar Tatineni, Tom Gibbs, John E Stone, Shantenu Jha, Arvind Ramanathan, Rommie E Amaro","doi":"10.1177/10943420211006452","DOIUrl":"https://doi.org/10.1177/10943420211006452","url":null,"abstract":"<p><p>We develop a generalizable AI-driven workflow that leverages heterogeneous HPC resources to explore the time-dependent dynamics of molecular systems. We use this workflow to investigate the mechanisms of infectivity of the SARS-CoV-2 spike protein, the main viral infection machinery. Our workflow enables more efficient investigation of spike dynamics in a variety of complex environments, including within a complete SARS-CoV-2 viral envelope simulation, which contains 305 million atoms and shows strong scaling on ORNL Summit using NAMD. We present several novel scientific discoveries, including the elucidation of the spike's full glycan shield, the role of spike glycans in modulating the infectivity of the virus, and the characterization of the flexible interactions between the spike and the human ACE2 receptor. We also demonstrate how AI can accelerate conformational sampling across different systems and pave the way for the future application of such methods to additional studies in SARS-CoV-2 and other molecular systems.</p>","PeriodicalId":54957,"journal":{"name":"International Journal of High Performance Computing Applications","volume":"35 5","pages":"432-451"},"PeriodicalIF":3.1,"publicationDate":"2021-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8064023/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140860617","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Special Issue Introduction: The Gordon Bell Special Prize for HPC-Based COVID-19 Research Finalists 特刊简介:戈登·贝尔基于hpc的COVID-19研究入围者特别奖
IF 3.1 3区 计算机科学 Q2 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2021-09-01 DOI: 10.1177/10943420211044760
B. Supinski
As the entire world realizes, 2020 was an extraordinary year. One of the brighter aspects of the year’s exceptionalism was the numerous demonstrations that high-performance computing (HPC) could contribute to solutions ofmany of themost difficult problems that our society faces. In recognition of those benefits for the one of the most pressing problems of 2020 and 2021, Gordon Bell, a pioneer in high-performance and parallel computing, endowed the Gordon Bell Special Prize for HPCBased COVID-19 Research. The prize recognizes outstanding research achievement towards the understanding of the COVID-19 pandemic through the use of HPC. The purpose of the award is to recognize the innovative parallel computing contributions towards the solution of the global crisis. This special issue presents the four papers that were selected as finalists for the award. The Gordon Bell Prize Committee selected these nominations based on performance and innovation in their computational methods, in addition to their contributions towards understanding the nature, spread and/or treatment of the disease. More specifically, the committee evaluated nominations on the basis of the following considerations:
{"title":"Special Issue Introduction: The Gordon Bell Special Prize for HPC-Based COVID-19 Research Finalists","authors":"B. Supinski","doi":"10.1177/10943420211044760","DOIUrl":"https://doi.org/10.1177/10943420211044760","url":null,"abstract":"As the entire world realizes, 2020 was an extraordinary year. One of the brighter aspects of the year’s exceptionalism was the numerous demonstrations that high-performance computing (HPC) could contribute to solutions ofmany of themost difficult problems that our society faces. In recognition of those benefits for the one of the most pressing problems of 2020 and 2021, Gordon Bell, a pioneer in high-performance and parallel computing, endowed the Gordon Bell Special Prize for HPCBased COVID-19 Research. The prize recognizes outstanding research achievement towards the understanding of the COVID-19 pandemic through the use of HPC. The purpose of the award is to recognize the innovative parallel computing contributions towards the solution of the global crisis. This special issue presents the four papers that were selected as finalists for the award. The Gordon Bell Prize Committee selected these nominations based on performance and innovation in their computational methods, in addition to their contributions towards understanding the nature, spread and/or treatment of the disease. More specifically, the committee evaluated nominations on the basis of the following considerations:","PeriodicalId":54957,"journal":{"name":"International Journal of High Performance Computing Applications","volume":"35 1","pages":"431 - 431"},"PeriodicalIF":3.1,"publicationDate":"2021-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"65398856","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Journal of High Performance Computing Applications
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1