A. Nishida, Hisashi Kotakemori, Tamito Kajiyama, Akira Nukada
Recent progress of science and technology has made numerical simulation an important approach for studies in various fields. Although scalable and high performance numerical libraries on large scale computing resources are indispensable tools for handling various multiscale phenomena, few projects for integrating these numerical libraries have been reported. The object of this project is the development of a basic library of solutions and algorithms required for large scale scientific simulations, which have been developed separately in each fields, and its integration into a scalable software infrastructure. The components include a scalable iterative solvers library Lis, having a number of solvers, preconditioners, and matrix storage formats that are flexibly combinable, a fast Fourier transform library FFTSS for various superscalar architectures with SIMD instructions, which outperforms some vendor-provided FFT libraries, and a language- and computing environment-independent matrix computation framework SILC. We show some highlights of our achievements on leading high performance computers.
{"title":"Scalable software infrastructure project","authors":"A. Nishida, Hisashi Kotakemori, Tamito Kajiyama, Akira Nukada","doi":"10.1145/1188455.1188601","DOIUrl":"https://doi.org/10.1145/1188455.1188601","url":null,"abstract":"Recent progress of science and technology has made numerical simulation an important approach for studies in various fields. Although scalable and high performance numerical libraries on large scale computing resources are indispensable tools for handling various multiscale phenomena, few projects for integrating these numerical libraries have been reported. The object of this project is the development of a basic library of solutions and algorithms required for large scale scientific simulations, which have been developed separately in each fields, and its integration into a scalable software infrastructure. The components include a scalable iterative solvers library Lis, having a number of solvers, preconditioners, and matrix storage formats that are flexibly combinable, a fast Fourier transform library FFTSS for various superscalar architectures with SIMD instructions, which outperforms some vendor-provided FFT libraries, and a language- and computing environment-independent matrix computation framework SILC. We show some highlights of our achievements on leading high performance computers.","PeriodicalId":115940,"journal":{"name":"Proceedings of the 2006 ACM/IEEE conference on Supercomputing","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117220069","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Leopold Grinberg, S. Dong, J. Noble, A. Yakhot, G. Karniadakis, N. Karonis
The human arterial tree consists of a complex network of branching blood vessels leading from the heart to arterioles, capillaries, and venules - comprising the microcirculation. The numerical simulation of the blood flow in a single part of the human arterial tree requires hundreds of CPUs; a full human arterial tree will require thousands of CPUs. Nowadays, we can use geographically distributed supercomputers connected by a fast network to perform large-scale simulations.Nektar-G2 is the grid-enabled version of Nektar, software developed at Brown University, that allows to solve problems on geographically distributed supercomputers. The topology-aware feature of MPICH-G2 is utilized to enforce an efficient data distribution strategy. Multi-level message passing algorithms minimizes the inter-site communication. Our ultimate goal is to model blood flow interaction of different regions of the cardiovascular system and to establish a biomechanics gateway on the TeraGrid.During poster presentation we will present results of ongoing project.
{"title":"Human arterial tree simulation on TeraGrid","authors":"Leopold Grinberg, S. Dong, J. Noble, A. Yakhot, G. Karniadakis, N. Karonis","doi":"10.1145/1188455.1188613","DOIUrl":"https://doi.org/10.1145/1188455.1188613","url":null,"abstract":"The human arterial tree consists of a complex network of branching blood vessels leading from the heart to arterioles, capillaries, and venules - comprising the microcirculation. The numerical simulation of the blood flow in a single part of the human arterial tree requires hundreds of CPUs; a full human arterial tree will require thousands of CPUs. Nowadays, we can use geographically distributed supercomputers connected by a fast network to perform large-scale simulations.Nektar-G2 is the grid-enabled version of Nektar, software developed at Brown University, that allows to solve problems on geographically distributed supercomputers. The topology-aware feature of MPICH-G2 is utilized to enforce an efficient data distribution strategy. Multi-level message passing algorithms minimizes the inter-site communication. Our ultimate goal is to model blood flow interaction of different regions of the cardiovascular system and to establish a biomechanics gateway on the TeraGrid.During poster presentation we will present results of ongoing project.","PeriodicalId":115940,"journal":{"name":"Proceedings of the 2006 ACM/IEEE conference on Supercomputing","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134090473","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Cluster computing is a disruptive force that has quickly reshaped the HPC market and rapidly gained acceptance as a solution to an ever-increasing number of business and research problems in the data center. To sustain, and accelerate the growth of grid and related fields, we need a set of reference architectures, implementations and technology standards that will enable interoperable components and an ubiquitous computing experience.This panel will debut the initiative to establish standards to discuss and implement grid systems. Industry and academic representatives will debate current challenges in grid computing, why we need standards, how we should talk about grid, what grid will mean to the industry for the next ten years, and how grid is a spring board for academics to take HPC, virtualization, and distributed computing to new heights. During the Q&A period, attendees will have an opportunity to voice their concerns or support of initiatives addressed.
{"title":"What's inside the grid? a discussion of standards and the future of computing","authors":"Gary Tyreman, Mark Linesch, S. Wheat, Andre Hill","doi":"10.1145/1188455.1188529","DOIUrl":"https://doi.org/10.1145/1188455.1188529","url":null,"abstract":"Cluster computing is a disruptive force that has quickly reshaped the HPC market and rapidly gained acceptance as a solution to an ever-increasing number of business and research problems in the data center. To sustain, and accelerate the growth of grid and related fields, we need a set of reference architectures, implementations and technology standards that will enable interoperable components and an ubiquitous computing experience.This panel will debut the initiative to establish standards to discuss and implement grid systems. Industry and academic representatives will debate current challenges in grid computing, why we need standards, how we should talk about grid, what grid will mean to the industry for the next ten years, and how grid is a spring board for academics to take HPC, virtualization, and distributed computing to new heights. During the Q&A period, attendees will have an opportunity to voice their concerns or support of initiatives addressed.","PeriodicalId":115940,"journal":{"name":"Proceedings of the 2006 ACM/IEEE conference on Supercomputing","volume":"53 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133267870","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Cluster Storage and File System TechnologiesTo meet the demands of increasingly hungry cluster applications, cluster-based distributed storage technologies are now capable of delivering performance scaling 10's to 100's of GB/sec. This tutorial will examine current state-of-the-art high performance file systems and the underlying technologies employed to deliver scalable performance across a range of scientific and industrial applications.The first half of the tutorial provides an in-depth description of the core features common across most high-performance file systems; including details of datapath design, decoupled and scalable metadata operations, data layout techniques, failover techniques, scalable reconstruction, storage interfaces and security. The second half describes the design trade-offs found in both open-source and commercial solutions including Lustre, GPFS, Parallel NFS and Panasas.
{"title":"Cluster storage and file system technologies","authors":"B. Welch, M. Unangst","doi":"10.1145/1188455.1188678","DOIUrl":"https://doi.org/10.1145/1188455.1188678","url":null,"abstract":"Cluster Storage and File System TechnologiesTo meet the demands of increasingly hungry cluster applications, cluster-based distributed storage technologies are now capable of delivering performance scaling 10's to 100's of GB/sec. This tutorial will examine current state-of-the-art high performance file systems and the underlying technologies employed to deliver scalable performance across a range of scientific and industrial applications.The first half of the tutorial provides an in-depth description of the core features common across most high-performance file systems; including details of datapath design, decoupled and scalable metadata operations, data layout techniques, failover techniques, scalable reconstruction, storage interfaces and security. The second half describes the design trade-offs found in both open-source and commercial solutions including Lustre, GPFS, Parallel NFS and Panasas.","PeriodicalId":115940,"journal":{"name":"Proceedings of the 2006 ACM/IEEE conference on Supercomputing","volume":"79 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133425013","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
I will describe and demonstrate the "Meeting List Tool", a shared application built with the Access Grid Toolkit for use in AG meetings.At present, the two most common ways of sharing text-based information during an AG meeting are shared presentations and the chat tool built into the Venue Client. The former is ideal for static content which is known in advance. The latter is ideal for sharing short pieces of information, such as a URL. What is apparently missing is an application for which data can be prepared in advance, displayed and quickly manipulated during a meeting, and kept at the close of the meeting by all the participants in the collaboration.This application is designed to fill this gap. The current version provides a list of items that can be highlighted, added, deleted, edited, and re-ordered, with the changes being propagated to all instances of the tool.
{"title":"The meeting list tool - a shared application for sharing dynamic information in meetings","authors":"Adam C. Carter","doi":"10.1145/1188455.1188786","DOIUrl":"https://doi.org/10.1145/1188455.1188786","url":null,"abstract":"I will describe and demonstrate the \"Meeting List Tool\", a shared application built with the Access Grid Toolkit for use in AG meetings.At present, the two most common ways of sharing text-based information during an AG meeting are shared presentations and the chat tool built into the Venue Client. The former is ideal for static content which is known in advance. The latter is ideal for sharing short pieces of information, such as a URL. What is apparently missing is an application for which data can be prepared in advance, displayed and quickly manipulated during a meeting, and kept at the close of the meeting by all the participants in the collaboration.This application is designed to fill this gap. The current version provides a list of items that can be highlighted, added, deleted, edited, and re-ordered, with the changes being propagated to all instances of the tool.","PeriodicalId":115940,"journal":{"name":"Proceedings of the 2006 ACM/IEEE conference on Supercomputing","volume":"169 ","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"113991445","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
M. Slawinska, Dawid Kurzyniec, Jaroslaw Slawinski, V. Sunderam
Shared HPC platforms continue to require substantial effort for software installation and management, often necessitating manual intervention and tedious procedures. We propose a novel model of resource sharing that shifts resource virtualization and aggregation responsibilities to client-side software, thus reducing the burdens on resource providers.The Zero-Force MPI toolkit automates the installation, build, run, and post-processing stages of HPC applications, thus allowing application scientists to focus on using resources instead of managing them. Through a provided console, MPI runtime systems, support libraries, application executables, and needed datafiles can be soft-installed across distributed resources with just a few commands. Built-in data synchronization capabilities simplify common HPC development tasks, saving end-user time and effort. To evaluate ZF-MPI, we conducted experiments with the NAS Parallel Benchmarks. Results demonstrate that the proposed run-not-install approach is effective and may substantially increase overall productivity.
{"title":"Zero-Force MPI: toward tractable toolkits for high performance computing","authors":"M. Slawinska, Dawid Kurzyniec, Jaroslaw Slawinski, V. Sunderam","doi":"10.1145/1188455.1188595","DOIUrl":"https://doi.org/10.1145/1188455.1188595","url":null,"abstract":"Shared HPC platforms continue to require substantial effort for software installation and management, often necessitating manual intervention and tedious procedures. We propose a novel model of resource sharing that shifts resource virtualization and aggregation responsibilities to client-side software, thus reducing the burdens on resource providers.The Zero-Force MPI toolkit automates the installation, build, run, and post-processing stages of HPC applications, thus allowing application scientists to focus on using resources instead of managing them. Through a provided console, MPI runtime systems, support libraries, application executables, and needed datafiles can be soft-installed across distributed resources with just a few commands. Built-in data synchronization capabilities simplify common HPC development tasks, saving end-user time and effort. To evaluate ZF-MPI, we conducted experiments with the NAS Parallel Benchmarks. Results demonstrate that the proposed run-not-install approach is effective and may substantially increase overall productivity.","PeriodicalId":115940,"journal":{"name":"Proceedings of the 2006 ACM/IEEE conference on Supercomputing","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122354528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
TotalView is a flexible, scriptable parallel debugger with wide acceptance in the High Performance Computing community. This BOF will be an opportunity for TotalView users to share clever and interesting ways of adapting TotalView to their unique environment, using TotalView to do something unusual, or simply making the day to day process of debugging easier. Contact Chris.Gottbrath@etnus.com if you want us to reserve time for you to tell your story or simply show up at the BOF and step forward.
{"title":"TotalView tips and tricks","authors":"C. Gottbrath, P. Thompson","doi":"10.1145/1188455.1188465","DOIUrl":"https://doi.org/10.1145/1188455.1188465","url":null,"abstract":"TotalView is a flexible, scriptable parallel debugger with wide acceptance in the High Performance Computing community. This BOF will be an opportunity for TotalView users to share clever and interesting ways of adapting TotalView to their unique environment, using TotalView to do something unusual, or simply making the day to day process of debugging easier. Contact Chris.Gottbrath@etnus.com if you want us to reserve time for you to tell your story or simply show up at the BOF and step forward.","PeriodicalId":115940,"journal":{"name":"Proceedings of the 2006 ACM/IEEE conference on Supercomputing","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129211183","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
F. Belletti, M. Cotallo, A. Flor, L. A. Fernández, A. Gordillo, A. Maiorano, F. Mantovani, E. Marinari, V. Martin-Mayor, A. M. Sudupe, D. Navarro, S. P. Gaviro, M. Rossi, J. Ruiz-Lorenzo, S. Schifano, D. Sciretti, A. Tarancón, R. Tripiccione, J. Velasco
IANUS is a massively parallel system based on a 2D array of FPGA-based processors with nearest-neighbor connections. Processors are also directly connected to a central hub attached to a host computer.The prototype, available in October 2006 uses an array of 4x4 Xilinx Virtex4LX160 FPGA's.We map onto the array the computational kernels of scientific applications characterized by regular control flow, unconventional mix of data-manipulation operations and limited memory usage.Careful VHDL coding of the kernel algorithms relevant for Monte Carlo simulation of spin-glass systems (our first application) yields impressive performances: single processor tests concurrently update ~1000 spins, so average spin-update time is 15 psec. This is ~60 times faster than accurately programmed 3,2 GHz PC's. We plan to build a 256 nodes system, roughly equivalent to 15000 PC's.This poster describes the architecture, the implementation and the methodology with which a specific application is mapped onto the system.
{"title":"IANUS: scientific computing on an FPGA-based architecture","authors":"F. Belletti, M. Cotallo, A. Flor, L. A. Fernández, A. Gordillo, A. Maiorano, F. Mantovani, E. Marinari, V. Martin-Mayor, A. M. Sudupe, D. Navarro, S. P. Gaviro, M. Rossi, J. Ruiz-Lorenzo, S. Schifano, D. Sciretti, A. Tarancón, R. Tripiccione, J. Velasco","doi":"10.1145/1188455.1188633","DOIUrl":"https://doi.org/10.1145/1188455.1188633","url":null,"abstract":"IANUS is a massively parallel system based on a 2D array of FPGA-based processors with nearest-neighbor connections. Processors are also directly connected to a central hub attached to a host computer.The prototype, available in October 2006 uses an array of 4x4 Xilinx Virtex4LX160 FPGA's.We map onto the array the computational kernels of scientific applications characterized by regular control flow, unconventional mix of data-manipulation operations and limited memory usage.Careful VHDL coding of the kernel algorithms relevant for Monte Carlo simulation of spin-glass systems (our first application) yields impressive performances: single processor tests concurrently update ~1000 spins, so average spin-update time is 15 psec. This is ~60 times faster than accurately programmed 3,2 GHz PC's. We plan to build a 256 nodes system, roughly equivalent to 15000 PC's.This poster describes the architecture, the implementation and the methodology with which a specific application is mapped onto the system.","PeriodicalId":115940,"journal":{"name":"Proceedings of the 2006 ACM/IEEE conference on Supercomputing","volume":"328 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129305455","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Simulations that model in detail complex interactions in complex environments are now possible, and visualization is a uniquely powerful analysis tool. However, visualization features of simulation codes are typically intended for scientists designing simulations, providing little support for presenting simulation results in a form suitable for non-experts. Conversely, graphics software and hardware progress has been fueled by applications where precisely abiding by the laws of physics is of secondary importance compared to visual realism.This half-day tutorial presents an approach for state-of-the-art visualization of simulation data based on connecting the worlds of computer simulation and computer animation. Concretely, the attendees will learn how to simplify, import, and integrate the simulation data into the surrounding scene, with examples from our simulation of the September 11 Attack on the Pentagon. The resulting realistic visualization enables effective dissemination of simulation results, and helps simulations reach their full potential for high societal impact.
{"title":"Realistic visualization for large-scale simulations","authors":"V. Popescu, C. Hoffmann","doi":"10.1145/1188455.1188688","DOIUrl":"https://doi.org/10.1145/1188455.1188688","url":null,"abstract":"Simulations that model in detail complex interactions in complex environments are now possible, and visualization is a uniquely powerful analysis tool. However, visualization features of simulation codes are typically intended for scientists designing simulations, providing little support for presenting simulation results in a form suitable for non-experts. Conversely, graphics software and hardware progress has been fueled by applications where precisely abiding by the laws of physics is of secondary importance compared to visual realism.This half-day tutorial presents an approach for state-of-the-art visualization of simulation data based on connecting the worlds of computer simulation and computer animation. Concretely, the attendees will learn how to simplify, import, and integrate the simulation data into the surrounding scene, with examples from our simulation of the September 11 Attack on the Pentagon. The resulting realistic visualization enables effective dissemination of simulation results, and helps simulations reach their full potential for high societal impact.","PeriodicalId":115940,"journal":{"name":"Proceedings of the 2006 ACM/IEEE conference on Supercomputing","volume":"44 24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128527539","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
H. Akiba, T. Ohyama, Y. Shibata, Kiyoshi Yuyama, Yoshikazu Katai, R. Takeuchi, T. Hoshino, S. Yoshimura, H. Noguchi, Manish Gupta, John A. Gunnels, V. Austel, Yogish Sabharwal, R. Garg, S. Kato, T. Kawakami, Satoru Todokoro, Junko Ikeda
Existing commercial finite element analysis (FEA) codes do not exhibit the performance necessary for large scale analysis on parallel computer systems. In this paper, we demonstrate the performance characteristics of a commercial parallel structural analysis code, ADVC, on Blue Gene/L (BG/L). The numerical algorithm of ADVC is described, tuned, and optimized on BG/L, and then a large scale drop impact analysis of a mobile phone is performed. The model of the mobile phone is a nearly-full assembly that includes inner structures. The size of the model we have analyzed has 47 million nodal points and 142 million DOFs. This does not seem exceptionally large, but the dynamic impact analysis of a product model, with the contact condition on the entire surface of the outer case under this size, cannot be handled by other CAE systems. Our analysis is an unprecedented attempt in the electronics industry. It took only half a day, 12.1 hours, for the analysis of about 2.4 milliseconds. The floating point operation performance obtained has been 538 GFLOPS on 4096 node of BG/L.
{"title":"Large scale drop impact analysis of mobile phone using ADVC on Blue Gene/L","authors":"H. Akiba, T. Ohyama, Y. Shibata, Kiyoshi Yuyama, Yoshikazu Katai, R. Takeuchi, T. Hoshino, S. Yoshimura, H. Noguchi, Manish Gupta, John A. Gunnels, V. Austel, Yogish Sabharwal, R. Garg, S. Kato, T. Kawakami, Satoru Todokoro, Junko Ikeda","doi":"10.1145/1188455.1188503","DOIUrl":"https://doi.org/10.1145/1188455.1188503","url":null,"abstract":"Existing commercial finite element analysis (FEA) codes do not exhibit the performance necessary for large scale analysis on parallel computer systems. In this paper, we demonstrate the performance characteristics of a commercial parallel structural analysis code, ADVC, on Blue Gene/L (BG/L). The numerical algorithm of ADVC is described, tuned, and optimized on BG/L, and then a large scale drop impact analysis of a mobile phone is performed. The model of the mobile phone is a nearly-full assembly that includes inner structures. The size of the model we have analyzed has 47 million nodal points and 142 million DOFs. This does not seem exceptionally large, but the dynamic impact analysis of a product model, with the contact condition on the entire surface of the outer case under this size, cannot be handled by other CAE systems. Our analysis is an unprecedented attempt in the electronics industry. It took only half a day, 12.1 hours, for the analysis of about 2.4 milliseconds. The floating point operation performance obtained has been 538 GFLOPS on 4096 node of BG/L.","PeriodicalId":115940,"journal":{"name":"Proceedings of the 2006 ACM/IEEE conference on Supercomputing","volume":"110 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128541567","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}