首页 > 最新文献

IEEE Parallel & Distributed Technology: Systems & Applications最新文献

英文 中文
Application development for distributed environments [Book Reviews] 分布式环境下的应用程序开发[书评]
Pub Date : 1900-01-01 DOI: 10.1109/M-PDT.1995.414846
M. Machura
This is the second book in the James Martin/ McGraw-Hill Productivity Series aimed a t information systems professionals and managers. Incidentally , the first book was written by the same author and is called Client/Servei-Computing. The series focuses on current computing technologies in an attempt to meet new challenges that modern organizations face. I was attracted to the book by its title. With hindsight, however, I think that a more appropriate title would be " Dezielopment Issues and Tools in Distributed Systewts. " T h e book provides a comprehensive picture of all the major elements of distributed systems as of early 1993. The author takes a pragmatic approach by concentrating on prevailing technologies, such as relational databases, structured design methods, cliendserver architectures, 4GLs and GUI builders. Dewire also pays due attention to the available standards. Roughly two thirds of the book contains a general discussion of distributed systems; the remaining one third surveys various development tools. Part 1 presents the basic concepts, application development strategies, and components of distributed systems. Part 2 deals with analysis and top-level design, and Part 3 covers the construction of distributed systems (detailed design and implementation). Part 4, called " Operations , " contains a chapter on integration that surveys the important issues of transaction management , , .work management, and distributed computing environments. This section also has a chapter on production that discusses configuration and version control, sharing data, monitoring networks, and security. Part 5 presents commercial application development products for distributed systems: 4GLs, cliendserver tools, and CASE tools. The concluding chapter discusses future trends. Application Development far Distributed Envi-m w " s stresses the importance of distributed, enterprise-wide information technology solutions in modern organizations that need to quickly respond to market changes and modify their business processes. Dewire estimates that 20% of the current distributed applications are mission-critical systems such as transaction-based operational 1%. The remaining 80% are less critical systems, such as information and decision support systems. CASE tools service the first category ; cliendserver development tools and 4GLs service the latter. As the technology matures, the cliendserver tools and CASE tools will merge, and 4GLs will evolve into flexible and efficient tools for cliendserver applications. As I mentioned earlier, the book covers the established distributed computing technologies. Dewire gives a rather careful, though insufficient, treatment to the emerging technologies such as distributed object computing. In fact, Dewire refrains from endorsing …
这是James Martin/ McGraw-Hill生产力系列的第二本书,面向信息系统专业人员和管理人员。顺便提一下,第一本书是由同一作者写的,名为“客户机/服务计算”。本系列着重于当前的计算技术,试图应对现代组织所面临的新挑战。我被书名吸引住了。然而,事后看来,我认为更合适的标题应该是“分布式系统中的开发问题和工具”。这本书提供了1993年初分布式系统所有主要元素的全面图景。作者采取了务实的方法,专注于流行的技术,如关系数据库,结构化设计方法,客户端服务器架构,4gl和GUI构建器。Dewire也对现有的标准给予了应有的关注。书中大约三分之二的内容是关于分布式系统的一般性讨论;剩下的三分之一调查了各种开发工具。第1部分介绍了分布式系统的基本概念、应用程序开发策略和组件。第2部分讨论分析和顶层设计,第3部分介绍分布式系统的构造(详细设计和实现)。第4部分名为“操作”,包含了关于集成的一章,该章调查了事务管理、工作管理和分布式计算环境的重要问题。本节还有一章是关于产品的,讨论配置和版本控制、共享数据、监视网络和安全性。第5部分介绍了用于分布式系统的商业应用程序开发产品:4gl、客户机服务器工具和CASE工具。最后一章讨论了未来的趋势。分布式环境的应用开发强调了分布式的、企业范围的信息技术解决方案在需要快速响应市场变化和修改其业务流程的现代组织中的重要性。Dewire估计,当前的分布式应用程序中有20%是任务关键型系统,例如基于事务的操作系统占1%。剩下的80%是不那么关键的系统,如信息和决策支持系统。CASE工具服务的第一类;客户端服务器开发工具和4gl服务于后者。随着技术的成熟,客户端服务器工具和CASE工具将会合并,4gl将演变为灵活高效的客户端服务器应用程序工具。正如我前面提到的,本书涵盖了已建立的分布式计算技术。Dewire对分布式对象计算等新兴技术进行了相当谨慎(尽管不够充分)的处理。事实上,德威尔不支持…
{"title":"Application development for distributed environments [Book Reviews]","authors":"M. Machura","doi":"10.1109/M-PDT.1995.414846","DOIUrl":"https://doi.org/10.1109/M-PDT.1995.414846","url":null,"abstract":"This is the second book in the James Martin/ McGraw-Hill Productivity Series aimed a t information systems professionals and managers. Incidentally , the first book was written by the same author and is called Client/Servei-Computing. The series focuses on current computing technologies in an attempt to meet new challenges that modern organizations face. I was attracted to the book by its title. With hindsight, however, I think that a more appropriate title would be \" Dezielopment Issues and Tools in Distributed Systewts. \" T h e book provides a comprehensive picture of all the major elements of distributed systems as of early 1993. The author takes a pragmatic approach by concentrating on prevailing technologies, such as relational databases, structured design methods, cliendserver architectures, 4GLs and GUI builders. Dewire also pays due attention to the available standards. Roughly two thirds of the book contains a general discussion of distributed systems; the remaining one third surveys various development tools. Part 1 presents the basic concepts, application development strategies, and components of distributed systems. Part 2 deals with analysis and top-level design, and Part 3 covers the construction of distributed systems (detailed design and implementation). Part 4, called \" Operations , \" contains a chapter on integration that surveys the important issues of transaction management , , .work management, and distributed computing environments. This section also has a chapter on production that discusses configuration and version control, sharing data, monitoring networks, and security. Part 5 presents commercial application development products for distributed systems: 4GLs, cliendserver tools, and CASE tools. The concluding chapter discusses future trends. Application Development far Distributed Envi-m w \" s stresses the importance of distributed, enterprise-wide information technology solutions in modern organizations that need to quickly respond to market changes and modify their business processes. Dewire estimates that 20% of the current distributed applications are mission-critical systems such as transaction-based operational 1%. The remaining 80% are less critical systems, such as information and decision support systems. CASE tools service the first category ; cliendserver development tools and 4GLs service the latter. As the technology matures, the cliendserver tools and CASE tools will merge, and 4GLs will evolve into flexible and efficient tools for cliendserver applications. As I mentioned earlier, the book covers the established distributed computing technologies. Dewire gives a rather careful, though insufficient, treatment to the emerging technologies such as distributed object computing. In fact, Dewire refrains from endorsing …","PeriodicalId":325213,"journal":{"name":"IEEE Parallel & Distributed Technology: Systems & Applications","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129484003","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Applied Parallel Research's xHPF system xHPF系统的应用并行研究
Pub Date : 1900-01-01 DOI: 10.1109/M-PDT.1994.329805
J. Levesque
Applied Parallel Research takes a somewhat different approach to High Performance Fortran than do other implementors. APR feels the real power of HPF is in its comment line directives by which the user can drive an automatic parallelization system. Rather than treating HPF as an altemative to automatic parallelization, we believe that it can be a powerful aid for automatic parallelization of existing Fortran 77 programs. W e have arrived at this point of view after a considerable effort to provide source-code global analyzers and parallelizers with extensive capabilities for large, real-world, sequential Fortran 77 programs. For example, our xHPF system will parallelize very complex Fortran 77 DO loops rather than relying on the user to explicitly expose parallel operations by translating to Fortran 90 array syntax. HPF’s data-distribution directives let us provide batch automatic parallelization tools, such as xHPF, in contrast to our interactive Forge 90 Distributed-Memory Parallelizer, which requires the user to explicitly direct the data decomposition of the arrays in the program. xHPF also accepts Fortran 90 array syntax and extends HPF data-distribution rules. APRs approach has been to provide HPF compilation systems that let users more easily port existing sequential Fortran 77 programs to MPP systems. APR feels that the market for tools that port existing Fortran programs to MPP systems far exceeds the market for tools to develop parallel programs from scratch.
应用并行研究对高性能Fortran采用了一种与其他实现者不同的方法。APR认为HPF的真正力量在于它的注释行指令,用户可以通过它来驱动自动并行化系统。我们不认为HPF是自动并行化的替代方案,而是认为它可以成为现有Fortran 77程序自动并行化的有力辅助。在为大型的、真实的、顺序的Fortran 77程序提供具有广泛功能的源代码全局分析器和并行化器的大量工作之后,我们得出了这个观点。例如,我们的xHPF系统将并行处理非常复杂的Fortran 77 DO循环,而不是依赖于用户通过转换为Fortran 90数组语法来显式地暴露并行操作。HPF的数据分布指令允许我们提供批处理自动并行化工具,如xHPF,而不是交互式的Forge 90分布式内存并行化工具,后者需要用户显式地指导程序中数组的数据分解。xHPF还接受Fortran 90数组语法并扩展了HPF数据分布规则。APRs的方法是提供HPF编译系统,让用户更容易地将现有的顺序Fortran 77程序移植到MPP系统。APR认为,将现有Fortran程序移植到MPP系统的工具市场远远超过从头开始开发并行程序的工具市场。
{"title":"Applied Parallel Research's xHPF system","authors":"J. Levesque","doi":"10.1109/M-PDT.1994.329805","DOIUrl":"https://doi.org/10.1109/M-PDT.1994.329805","url":null,"abstract":"Applied Parallel Research takes a somewhat different approach to High Performance Fortran than do other implementors. APR feels the real power of HPF is in its comment line directives by which the user can drive an automatic parallelization system. Rather than treating HPF as an altemative to automatic parallelization, we believe that it can be a powerful aid for automatic parallelization of existing Fortran 77 programs. W e have arrived at this point of view after a considerable effort to provide source-code global analyzers and parallelizers with extensive capabilities for large, real-world, sequential Fortran 77 programs. For example, our xHPF system will parallelize very complex Fortran 77 DO loops rather than relying on the user to explicitly expose parallel operations by translating to Fortran 90 array syntax. HPF’s data-distribution directives let us provide batch automatic parallelization tools, such as xHPF, in contrast to our interactive Forge 90 Distributed-Memory Parallelizer, which requires the user to explicitly direct the data decomposition of the arrays in the program. xHPF also accepts Fortran 90 array syntax and extends HPF data-distribution rules. APRs approach has been to provide HPF compilation systems that let users more easily port existing sequential Fortran 77 programs to MPP systems. APR feels that the market for tools that port existing Fortran programs to MPP systems far exceeds the market for tools to develop parallel programs from scratch.","PeriodicalId":325213,"journal":{"name":"IEEE Parallel & Distributed Technology: Systems & Applications","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130208858","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
PGHPF from The Portland Group 波特兰集团的PGHPF
Pub Date : 1900-01-01 DOI: 10.1109/M-PDT.1994.329807
V. Schuster
PGHPF, The Portland Group’s HPF compiler, is now available for general distribution. Its initial release fully supports the HPF subset as defined in version 1 .O of the H P F Language Specification. A March 1995 release will support the full HPF language. PGHPF is available in two forms. A highly tuned version is integrated with PGI’s PGF77 Fortran compiler and produces executable images for most 8 6 0 and Sparc multiprocessor platforms. In this form, PGHPF will be the standard HPF compiler provided on the Intel Paragon and Meiko CS-2 scalable parallel processing systems. It will also be optimized for other 8 6 0 and SuperSparc sharedmemory multiprocessor systems. PGHPF is also available as a source-to-source translator that produces Fortran 77, incorporating calls to a portable communications library. This output, with linearized array references and de facto standard Cray pointer variable declarations, can then be used as input to standard node compilers. Both forms of the compiler use an internally defined transport-independent runtime library. This allows common source generation regardless of the target or the underlying communication mechanism (MPI, PVM, Parmacs, NX, or a targetcustom communication protocol). The runtime library for a specified target can thus be optimized outside the context of the compiler. PGI is developing optimized versions of the runtime library for the Intel Paragon, Meiko CS-2, SGI MP Challenge, SuperSparc workstation clusters, and Solaris shared-memory systems. Interfaces to PGHPF, including the runtime interface, will be open and freely available. This will let system vendors and researchers custom-tune for a specific target, and will facilitate integration with existing parallel support tools. The success of HPF as a standard depends on whether programmers can use it to implement efficient, portable versions of appropriate data-parallel applications. Based on that assumption, the highest priority for the initial release of PGHPF is completeness, correctness, and source portability. The initial release of PGHPF supports all of the HPF subset and will distribute and align data exactly as the programmer specifies, in as many dimensions as desired. Control parallelism will be exploited wherever possible as dictated by data distributions and language elements. PGI is spending significant effort to minimize the inefficiencies and overhead introduced to support the HPF paradigm. From a performance standpoint, minimization and efficiency of communication are most important. PGHPF incorporates optimizations that address both structured and unstructured communication. It can identify and exploit a program’s inherent structure through calls to structured asynchronous communication primitives. Examples of such primitives include collective shifts, the various forms of broadcast, and data reductions. Exploiting an application’s structure increases efficiency and performance portability. The asynchronous nature of the primitiv
PGHPF, Portland Group的HPF编译器,现在可以用于一般发行版。它的初始版本完全支持HPF语言规范1.0版中定义的HPF子集。1995年3月的版本将支持完整的HPF语言。PGHPF有两种形式。一个高度调优的版本与PGI的PGF77 Fortran编译器集成,并为大多数860和Sparc多处理器平台生成可执行映像。在这种形式下,PGHPF将成为Intel Paragon和Meiko CS-2可扩展并行处理系统上提供的标准HPF编译器。它还将针对其他860和SuperSparc共享内存多处理器系统进行优化。PGHPF还可以作为生成Fortran 77的源到源转换器,包含对便携式通信库的调用。该输出带有线性化的数组引用和事实上的标准Cray指针变量声明,然后可以用作标准节点编译器的输入。两种形式的编译器都使用内部定义的与传输无关的运行时库。这允许生成公共源代码,而不考虑目标或底层通信机制(MPI、PVM、Parmacs、NX或目标自定义通信协议)。因此,可以在编译器上下文之外对指定目标的运行时库进行优化。PGI正在为Intel Paragon、Meiko CS-2、SGI MP Challenge、SuperSparc工作站集群和Solaris共享内存系统开发优化版本的运行时库。PGHPF的接口,包括运行时接口,将是开放和免费提供的。这将允许系统供应商和研究人员针对特定目标进行定制调整,并将促进与现有并行支持工具的集成。HPF作为标准的成功取决于程序员是否可以使用它来实现适当的数据并行应用程序的高效、可移植版本。基于这一假设,PGHPF初始版本的最高优先级是完整性、正确性和源代码可移植性。PGHPF的初始版本支持所有HPF子集,并将按照程序员的要求在尽可能多的维度上精确地分发和对齐数据。控制并行性将根据数据分布和语言元素的要求尽可能加以利用。PGI正在花费大量精力来最小化为支持HPF范式而引入的低效率和开销。从性能的角度来看,沟通的最小化和效率是最重要的。PGHPF结合了处理结构化和非结构化通信的优化。它可以通过调用结构化异步通信原语来识别和利用程序的固有结构。这类原语的例子包括集体转换、各种形式的广播和数据缩减。利用应用程序的结构可以提高效率和性能可移植性。原语的异步特性允许通信与计算重叠,并且可以减少或消除某些应用程序中的通信配置文件。此外,这种方法允许许多特定于通信的优化,包括公共通信消除、通信调度、通信矢量化和非结构化通信调度信息的重用。应该在PGHPF下表现良好的HPF程序包括那些具有适合目标体系结构的显式数据分布和对齐的程序,以及那些自由使用FORALL、Fortran 9(数组赋值)和INDEPENDENT指令的程序。诸如此类的程序允许编译器基于程序员所表达的并行性进行优化。随着自动数据分布和并行化技术的发展,仔细的编码将变得不那么重要。在一台15节点的Intel Paragon上,PGHPF在Shallow Watei基准测试(9.3 ~)和3D Poisson Solvei(8.3倍)的原生HPF版本上表现出良好的加速效果。几个完整的应用程序也已经运行,包括16000线流体流动应用程序(6x)和3000线弹性波模拟应用程序(7 ~)。这些在共享内存SuperSparc系统上的测试显示了类似的可伸缩性。PGI认为这些效率数字是一个良好的开端,并且正在实现与目标无关的优化,这将进一步提高效率。HPF允许对应用程序固有的daE并行性进行简洁和可移植的规范。这是一种很有价值的方法,程序员可以通过它向编译器传达如何在存在内存层次结构的情况下进行最佳优化。PGI期望并行系统的程序员对访问在给定目标系统上最大限度地利用其信息的HPF编译器感兴趣,至少在短期内是这样。
{"title":"PGHPF from The Portland Group","authors":"V. Schuster","doi":"10.1109/M-PDT.1994.329807","DOIUrl":"https://doi.org/10.1109/M-PDT.1994.329807","url":null,"abstract":"PGHPF, The Portland Group’s HPF compiler, is now available for general distribution. Its initial release fully supports the HPF subset as defined in version 1 .O of the H P F Language Specification. A March 1995 release will support the full HPF language. PGHPF is available in two forms. A highly tuned version is integrated with PGI’s PGF77 Fortran compiler and produces executable images for most 8 6 0 and Sparc multiprocessor platforms. In this form, PGHPF will be the standard HPF compiler provided on the Intel Paragon and Meiko CS-2 scalable parallel processing systems. It will also be optimized for other 8 6 0 and SuperSparc sharedmemory multiprocessor systems. PGHPF is also available as a source-to-source translator that produces Fortran 77, incorporating calls to a portable communications library. This output, with linearized array references and de facto standard Cray pointer variable declarations, can then be used as input to standard node compilers. Both forms of the compiler use an internally defined transport-independent runtime library. This allows common source generation regardless of the target or the underlying communication mechanism (MPI, PVM, Parmacs, NX, or a targetcustom communication protocol). The runtime library for a specified target can thus be optimized outside the context of the compiler. PGI is developing optimized versions of the runtime library for the Intel Paragon, Meiko CS-2, SGI MP Challenge, SuperSparc workstation clusters, and Solaris shared-memory systems. Interfaces to PGHPF, including the runtime interface, will be open and freely available. This will let system vendors and researchers custom-tune for a specific target, and will facilitate integration with existing parallel support tools. The success of HPF as a standard depends on whether programmers can use it to implement efficient, portable versions of appropriate data-parallel applications. Based on that assumption, the highest priority for the initial release of PGHPF is completeness, correctness, and source portability. The initial release of PGHPF supports all of the HPF subset and will distribute and align data exactly as the programmer specifies, in as many dimensions as desired. Control parallelism will be exploited wherever possible as dictated by data distributions and language elements. PGI is spending significant effort to minimize the inefficiencies and overhead introduced to support the HPF paradigm. From a performance standpoint, minimization and efficiency of communication are most important. PGHPF incorporates optimizations that address both structured and unstructured communication. It can identify and exploit a program’s inherent structure through calls to structured asynchronous communication primitives. Examples of such primitives include collective shifts, the various forms of broadcast, and data reductions. Exploiting an application’s structure increases efficiency and performance portability. The asynchronous nature of the primitiv","PeriodicalId":325213,"journal":{"name":"IEEE Parallel & Distributed Technology: Systems & Applications","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116943150","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Alpha and oracle serve up very large memory [New Products] Alpha和oracle提供非常大的内存[新产品]
Pub Date : 1900-01-01 DOI: 10.1109/m-pdt.1995.414853
The Alphaserver 8400 enterprise server and Alphaserver 8200 departmental server use the 300-MHz Alpha 21164 chip, which can operate a t a billion instructions per second, according to Digital. The servers combine Alpha 64-bit architecture and very large memory capacity (up to 14 Gbytes). They offer a choice of PCI, XMI, and Futurebus+ buses. The AlphaServers have reliability and availability features such as OpenVMS clusters, hot swap disks, RAID, redundant power, ECC memory and data paths, fault management, and uninterruptible power system. They are available with Digital Unix or Open VMS operating systems. Digital also plans support for Windows NT. The Alphaserver 8200 features one to six processors and up to 6 Gbytes of memory. The base system costs $100,000. It includes one processor; power and packaging with a five-slot system bus for CPU, memory, and VO modules; 12 8 Mbytes of memory; an integrated VO module with SCSI and communication ports; a CDROM reader; and the OpenVMS or Digital Unix operating system. The Alphaserver 8400 features one to 12 processors and up to 14 Gbytes of memory. The base system, priced at $195,000, has the same basic configuration as the Alphaserver 8200 base system, but offers more expansion for additional CPU, memory, and VO connectivity, and twice the memory. T o support the AlphaServers, Oracle offers a very large memory option for its Oracle7 database. This option exploits the 64-bit Alpha architecture, Digital Unix, and the new server’s 14-Gbyte maximum main memory to allow a larger portion of the database to reside in memory. The option features two components: Large Systems Global Areas and Big Oracle Blocks. LSGAs are database buffer caches in excess of 2 Gbytes. According to Oracle, the LSGA is transparent to most applications, and application code does not have to be changed. BOBS support block sizes up to 32 Kbytes. Larger blocks allow more rows per block, meaning less overhead per row and fewer disk I/O requests when scanning tables, claims the company. Consequently, the database can move data from disk to memory and back much faster. Circle reader service number 23
据Digital称,Alphaserver 8400企业服务器和Alphaserver 8200部门服务器使用300兆赫的Alpha 21164芯片,每秒可以执行10亿条指令。服务器结合了Alpha 64位架构和非常大的内存容量(高达14 gb)。它们提供PCI、xml和Futurebus+总线的选择。alphaserver具有OpenVMS集群、热插拔盘、RAID、冗余电源、ECC内存和数据路径、故障管理、不间断电源系统等可靠性和可用性特性。它们可用于Digital Unix或Open VMS操作系统。Digital还计划支持Windows NT。Alphaserver 8200具有1到6个处理器和高达6gb的内存。基础系统的成本为10万美元。它包括一个处理器;电源和封装与一个五槽系统总线的CPU,内存和VO模块;128mb内存;具有SCSI和通信端口的集成VO模块;光盘读取器;和OpenVMS或数字Unix操作系统。Alphaserver 8400具有1到12个处理器和高达14gb的内存。基本系统的价格为19.5万美元,与Alphaserver 8200基本系统具有相同的基本配置,但提供了更多的扩展,可以提供额外的CPU、内存和VO连接,并且内存是其两倍。为了支持alphaserver, Oracle为其Oracle7数据库提供了一个非常大的内存选项。这个选项利用64位Alpha体系结构、Digital Unix和新服务器最大14gb的主存,允许更大的数据库部分驻留在内存中。该选项有两个组成部分:大系统全局区域和大Oracle块。lsga是超过2gb的数据库缓冲区缓存。根据Oracle的说法,LSGA对大多数应用程序是透明的,并且不需要更改应用程序代码。bob支持最大32 kb的块大小。该公司声称,更大的块允许每个块有更多的行,这意味着在扫描表时每行的开销更少,磁盘I/O请求更少。因此,数据库可以更快地将数据从磁盘移到内存并移回来。圈读者服务号码23
{"title":"Alpha and oracle serve up very large memory [New Products]","authors":"","doi":"10.1109/m-pdt.1995.414853","DOIUrl":"https://doi.org/10.1109/m-pdt.1995.414853","url":null,"abstract":"The Alphaserver 8400 enterprise server and Alphaserver 8200 departmental server use the 300-MHz Alpha 21164 chip, which can operate a t a billion instructions per second, according to Digital. The servers combine Alpha 64-bit architecture and very large memory capacity (up to 14 Gbytes). They offer a choice of PCI, XMI, and Futurebus+ buses. The AlphaServers have reliability and availability features such as OpenVMS clusters, hot swap disks, RAID, redundant power, ECC memory and data paths, fault management, and uninterruptible power system. They are available with Digital Unix or Open VMS operating systems. Digital also plans support for Windows NT. The Alphaserver 8200 features one to six processors and up to 6 Gbytes of memory. The base system costs $100,000. It includes one processor; power and packaging with a five-slot system bus for CPU, memory, and VO modules; 12 8 Mbytes of memory; an integrated VO module with SCSI and communication ports; a CDROM reader; and the OpenVMS or Digital Unix operating system. The Alphaserver 8400 features one to 12 processors and up to 14 Gbytes of memory. The base system, priced at $195,000, has the same basic configuration as the Alphaserver 8200 base system, but offers more expansion for additional CPU, memory, and VO connectivity, and twice the memory. T o support the AlphaServers, Oracle offers a very large memory option for its Oracle7 database. This option exploits the 64-bit Alpha architecture, Digital Unix, and the new server’s 14-Gbyte maximum main memory to allow a larger portion of the database to reside in memory. The option features two components: Large Systems Global Areas and Big Oracle Blocks. LSGAs are database buffer caches in excess of 2 Gbytes. According to Oracle, the LSGA is transparent to most applications, and application code does not have to be changed. BOBS support block sizes up to 32 Kbytes. Larger blocks allow more rows per block, meaning less overhead per row and fewer disk I/O requests when scanning tables, claims the company. Consequently, the database can move data from disk to memory and back much faster. Circle reader service number 23","PeriodicalId":325213,"journal":{"name":"IEEE Parallel & Distributed Technology: Systems & Applications","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115481602","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Portability and performance for parallel processing [Book Reviews] 并行处理的可移植性和性能[书评]
Pub Date : 1900-01-01 DOI: 10.1109/M-PDT.1995.414848
M. Paprzycki
Portability and Performance for Parallel Processing edited by Tony Hey and Jeanne Ferrante 272 pages $49.95 John Wiley &Sons, Chichester, UK 1994 ISBN 0-47 1-94246-4 retrieving data due to polling. This cost is negligible if the system is close to periodic, a somewhat uncommon situation in distributed environments. Schiitz then introduces the concept of testing a distributed system. He presents three ways to do a cluster test, where a cluster is a set of nodes forming a part of a distributed system:
并行处理的可移植性和性能由Tony Hey和Jeanne Ferrante编辑,272页$49.95 John Wiley &Sons,奇切斯特,英国1994 ISBN 0-47 1-94246-4由于轮询而检索数据。如果系统接近周期性,则此成本可以忽略不计,这在分布式环境中有些不常见。Schiitz随后介绍了测试分布式系统的概念。他提出了三种进行集群测试的方法,其中集群是组成分布式系统一部分的一组节点:
{"title":"Portability and performance for parallel processing [Book Reviews]","authors":"M. Paprzycki","doi":"10.1109/M-PDT.1995.414848","DOIUrl":"https://doi.org/10.1109/M-PDT.1995.414848","url":null,"abstract":"Portability and Performance for Parallel Processing edited by Tony Hey and Jeanne Ferrante 272 pages $49.95 John Wiley &Sons, Chichester, UK 1994 ISBN 0-47 1-94246-4 retrieving data due to polling. This cost is negligible if the system is close to periodic, a somewhat uncommon situation in distributed environments. Schiitz then introduces the concept of testing a distributed system. He presents three ways to do a cluster test, where a cluster is a set of nodes forming a part of a distributed system:","PeriodicalId":325213,"journal":{"name":"IEEE Parallel & Distributed Technology: Systems & Applications","volume":"57 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133895837","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Massively parallel artificial intelligence [Review] 大规模并行人工智能[回顾]
Pub Date : 1900-01-01 DOI: 10.1109/M-PDT.1995.414836
B. Mikolajczak
collection falters due to its lack of organization. Even though each paper addresses an important point related to efficient portable parallel computing and is worth reading in its own right, the collection remains just an assembly dispersed around a common subject. This is especially true for the last two chapters. In addition, the book fails to address one of the most popular (if not necessarily the best) attempts at providing software support for efficient portable parallel computing-the Parallel Virtual Machine (PVM) project from Oak Ridge National Laboratory. Nor is there any discussion of the current research from the High Performance Fortran (HPF) and MPI projects. Although these projects were not conference subjects , they are consequential and merit discussion. Having said all this, I must emphasize that this text is important; it explores one of the areas that are crucial for the success of parallel computing. When the editors prepared the book no one could foresee that a number of parallel computer vendors would go out of business or that a wave of strong criticism would be raised against the High Performance Computing Research Program. It is clear now that these occurrences grew out the continued lack of development environments for efficient portable programs, which raised doubts about the endeavor's commercial viability. In summary, this book will be of definite interest to anyone who has professional interest in parallel computing: computer scientists as well as engineers. It is a valuable resource that will introduce them to a variety of issues related to achieving efficient portable parallel computing. Each chapter contains an appropriate number of references that should allow further investigation. At the same time, the book certainly does not aspire to provide a complete overview of the field or give definitive answers. Since most of the papers require an overall understanding of parallel computing (some chapters go into considerable detail) the book is not pamcularly suitable as a textbook. However, this collection can function as a source of individual articles for use in the classroom or self-instruction. This collection comprises 12 papers devoted to different aspects of artificial intelligence as perceived , motivated, and applied by recent progress in massively parallel computer technology. The first paper by Kitano sets the stage for the following presentations, as it gives an overview of potential and real applications of massively parallel processing in artificial intelligence. The remaining papers are devoted to the following …
由于缺乏组织,募捐活动步履蹒跚。尽管每篇论文都阐述了与高效便携式并行计算相关的重要观点,并且值得一读,但该集合仍然只是围绕一个共同主题分散的汇编。最后两章尤其如此。此外,本书没有提到为高效便携式并行计算提供软件支持的最流行(如果不一定是最好的)尝试之一——橡树岭国家实验室的并行虚拟机(PVM)项目。也没有任何关于高性能Fortran (HPF)和MPI项目的当前研究的讨论。虽然这些项目不是会议的主题,但它们是重要的,值得讨论。说了这么多,我必须强调这一案文很重要;它探讨了对并行计算的成功至关重要的一个领域。当编辑们准备这本书的时候,没有人能预见到许多并行计算机供应商将会破产,或者对高性能计算研究计划的强烈批评浪潮将会兴起。现在很清楚,这些事件的发生是由于持续缺乏有效的可移植程序的开发环境,这引起了对努力的商业可行性的怀疑。总之,这本书肯定会对任何对并行计算有专业兴趣的人感兴趣:计算机科学家以及工程师。这是一个有价值的资源,它将向他们介绍与实现高效可移植并行计算相关的各种问题。每章都包含适当数量的参考资料,以便进一步研究。与此同时,本书当然不希望提供该领域的完整概述或给出明确的答案。由于大多数论文都要求对并行计算有一个全面的了解(有些章节非常详细),这本书不是特别适合作为教科书。然而,这个集合可以作为单个文章的来源,用于课堂或自学。本合集包括12篇论文,致力于通过大规模并行计算机技术的最新进展来感知、激发和应用人工智能的不同方面。Kitano的第一篇论文为接下来的演讲奠定了基础,因为它概述了大规模并行处理在人工智能中的潜在和实际应用。其余的论文致力于以下方面……
{"title":"Massively parallel artificial intelligence [Review]","authors":"B. Mikolajczak","doi":"10.1109/M-PDT.1995.414836","DOIUrl":"https://doi.org/10.1109/M-PDT.1995.414836","url":null,"abstract":"collection falters due to its lack of organization. Even though each paper addresses an important point related to efficient portable parallel computing and is worth reading in its own right, the collection remains just an assembly dispersed around a common subject. This is especially true for the last two chapters. In addition, the book fails to address one of the most popular (if not necessarily the best) attempts at providing software support for efficient portable parallel computing-the Parallel Virtual Machine (PVM) project from Oak Ridge National Laboratory. Nor is there any discussion of the current research from the High Performance Fortran (HPF) and MPI projects. Although these projects were not conference subjects , they are consequential and merit discussion. Having said all this, I must emphasize that this text is important; it explores one of the areas that are crucial for the success of parallel computing. When the editors prepared the book no one could foresee that a number of parallel computer vendors would go out of business or that a wave of strong criticism would be raised against the High Performance Computing Research Program. It is clear now that these occurrences grew out the continued lack of development environments for efficient portable programs, which raised doubts about the endeavor's commercial viability. In summary, this book will be of definite interest to anyone who has professional interest in parallel computing: computer scientists as well as engineers. It is a valuable resource that will introduce them to a variety of issues related to achieving efficient portable parallel computing. Each chapter contains an appropriate number of references that should allow further investigation. At the same time, the book certainly does not aspire to provide a complete overview of the field or give definitive answers. Since most of the papers require an overall understanding of parallel computing (some chapters go into considerable detail) the book is not pamcularly suitable as a textbook. However, this collection can function as a source of individual articles for use in the classroom or self-instruction. This collection comprises 12 papers devoted to different aspects of artificial intelligence as perceived , motivated, and applied by recent progress in massively parallel computer technology. The first paper by Kitano sets the stage for the following presentations, as it gives an overview of potential and real applications of massively parallel processing in artificial intelligence. The remaining papers are devoted to the following …","PeriodicalId":325213,"journal":{"name":"IEEE Parallel & Distributed Technology: Systems & Applications","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130321220","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pacific Sierra's VAST-HPF and VAST/77toHPF 太平洋山脉的VAST- hpf和VAST/77toHPF
Pub Date : 1900-01-01 DOI: 10.1109/M-PDT.1994.329809
J. Vanderlip
WHAT CLASS OF HPF PROGRAMS WILL PERFORM WELL? To perform well, HPF programs must spend almost all their time in sections of code that can be partitioned across the processors. They also must access data that resides on the local processor almost all the time, and send or receive data from other processors very infrequently. This means that an HPF program should spend its time almost completely in array syntax or loops that can be performed in parallel, and should be written so that references to arrays in loops are aligned and distributed in the same way. VAST-HPF performs well on shifted sections. Real programs often use sections of arrays that are offset in one or more dimensions. A common construct in grid-based computations is the use of slightly shifted sections of arrays in nearestneighbor computations. For blockdistributed arrays, this means that communication is needed at the boundaries of the blocks. VAST-HPF makes the local distribution of such arrays slightly larger so that the edge values can be communicated into this expanded region. It enhances data locality by passing messages only for the elements at the edge of the offset section. VAST-HPF also performs well on reductions. Reduction operations, such as the summation of array elements, occur frequently in real programs. VAST-HPF handles reductions by reducing the distributed part of the array calculation on each processor, passing the partial reductions to a single processor for the final reduction, and then broadcasting the final result to all processors.
哪一类的HPF程序将表现良好?为了更好地执行,HPF程序必须把几乎所有的时间都花在可以跨处理器分区的代码段上。它们还必须几乎一直访问驻留在本地处理器上的数据,并且很少从其他处理器发送或接收数据。这意味着HPF程序应该几乎完全把时间花在可以并行执行的数组语法或循环上,并且应该在编写时使循环中对数组的引用以相同的方式对齐和分布。VAST-HPF在移位截面上表现良好。实际的程序经常使用在一个或多个维度上偏移的数组段。在基于网格的计算中,一个常见的构造是在最近邻计算中使用稍微移位的数组部分。对于块分布式数组,这意味着需要在块的边界进行通信。瓦斯- hpf使这种阵列的局部分布稍微大一些,以便边缘值可以传递到这个扩展的区域。它通过仅为偏移部分边缘的元素传递消息来增强数据局部性。VAST-HPF在缩减方面也表现良好。简化操作,如数组元素的求和,在实际程序中经常发生。VAST-HPF通过减少每个处理器上的分布式数组计算部分来处理缩减,将部分缩减传递给单个处理器以进行最终缩减,然后将最终结果广播给所有处理器。
{"title":"Pacific Sierra's VAST-HPF and VAST/77toHPF","authors":"J. Vanderlip","doi":"10.1109/M-PDT.1994.329809","DOIUrl":"https://doi.org/10.1109/M-PDT.1994.329809","url":null,"abstract":"WHAT CLASS OF HPF PROGRAMS WILL PERFORM WELL? To perform well, HPF programs must spend almost all their time in sections of code that can be partitioned across the processors. They also must access data that resides on the local processor almost all the time, and send or receive data from other processors very infrequently. This means that an HPF program should spend its time almost completely in array syntax or loops that can be performed in parallel, and should be written so that references to arrays in loops are aligned and distributed in the same way. VAST-HPF performs well on shifted sections. Real programs often use sections of arrays that are offset in one or more dimensions. A common construct in grid-based computations is the use of slightly shifted sections of arrays in nearestneighbor computations. For blockdistributed arrays, this means that communication is needed at the boundaries of the blocks. VAST-HPF makes the local distribution of such arrays slightly larger so that the edge values can be communicated into this expanded region. It enhances data locality by passing messages only for the elements at the edge of the offset section. VAST-HPF also performs well on reductions. Reduction operations, such as the summation of array elements, occur frequently in real programs. VAST-HPF handles reductions by reducing the distributed part of the array calculation on each processor, passing the partial reductions to a single processor for the final reduction, and then broadcasting the final result to all processors.","PeriodicalId":325213,"journal":{"name":"IEEE Parallel & Distributed Technology: Systems & Applications","volume":"90 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114245221","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
UI-based design and development for client/server applications [Book Reviews] 基于ui的客户端/服务器应用程序设计与开发[书评]
Pub Date : 1900-01-01 DOI: 10.1109/M-PDT.1995.414849
M. Trayton
In this text the authors state that they use the top 4GL integrated development environment products “to illustrate the development of a large-scale client/server application example.” Beginning very enthusiastically, the authors talk about the book as if it is the outcome of a very successful project. The authors intend the audience to be mainframe professionals who need help translating their skills for client-server, object-oriented, and graphical user interface applications (I abbreviate these terms as CS/OO/GUI). The book starts by using abbreviations that only computer professionals with a mainframe background would recognize. As the book moves on, CS/OO/GUI jargon becomes commonplace. The authors should have included a glossary. As it is, the reader must continually review previous pages to find the meaning of acronyms and abbreviations. This is annoying and it becomes clear that the reader needs quite a lot of knowledge of CS/OO/GUI to understand what the authors are saying. The authors make the assumption that, for the foreseeable future, CS/oO/GUI is the way to go in data processing and that the important new skills involve the “new wave” of CS/OO/GUI 4GL workbenches, such as PowerBuilder, Visual Basic, SQLWindows, and PARTS Workbench. The book includes discussions on the pros and cons of corporate mainframes, centralized data processing, multiple parallel processor machines, midrange minicomputers, workstations, and personal computers (“The key to CS”). The text then goes on to explain the different operating systems related to each type of hardware platform and their possibilities for the future. The authors consider relational databases to be the “cornerstone” of CS computing. They suggest that an insight into computer communication is critical in understanding CS. There then follows a general overview of computer communications and the background to CS systems. After an overview of CS/OO/GUI (this can be skipped by the more enlightened reader), the authors build a sample business application with each of the four products. The example application that the authors use is small enough to be built in a short amount of time by one person. The book emphasizes the importance of good GUI design, recommends adhering to GUI standards, encourages the use of meaningful variable names, and stresses the importance of building on-line help into a system. The authors discuss the hardware and software requirements in considerable detail; this is useful for those thinking about buying one of the products mentioned in the book. When describing SQLWindows, the authors briefly mention project management, but only in relation to the facilities available for this in the product. The authors encourage interactive development, but unfortunately they approach development by prototyping and hacking the application together. The authors spend almost no time discussing the development of larger systems that require a project team. On first seeing the book, I t
在这篇文章中,作者声明他们使用了顶级的4GL集成开发环境产品“来说明大规模客户机/服务器应用程序示例的开发”。一开始,作者非常热情地谈论这本书,就好像它是一个非常成功的项目的结果。作者希望读者是大型机专业人员,他们需要帮助将他们的技能转化为客户机-服务器、面向对象和图形用户界面应用程序(我将这些术语缩写为CS/OO/GUI)。这本书从使用缩写开始,只有具有大型机背景的计算机专业人员才能识别。随着本书的深入,CS/OO/GUI术语变得司空见惯。作者应该包括一个术语表。事实上,读者必须不断地回顾前几页,以找出首字母缩写词的含义。这很烦人,很明显,读者需要相当多的CS/OO/GUI知识才能理解作者所说的话。作者假设,在可预见的未来,CS/oO/GUI是数据处理的发展方向,重要的新技能涉及CS/oO/GUI 4GL工作台的“新浪潮”,如PowerBuilder、Visual Basic、SQLWindows和PARTS Workbench。这本书包括对企业大型机、集中数据处理、多并行处理器机器、中型小型机、工作站和个人计算机(“计算机科学的关键”)的利弊的讨论。文章接着解释了与每种类型的硬件平台相关的不同操作系统及其未来的可能性。作者认为关系数据库是计算机科学计算的“基石”。他们认为,对计算机通信的洞察是理解计算机科学的关键。然后是计算机通信的总体概述和CS系统的背景。在概述CS/OO/GUI之后(更有见识的读者可以跳过这部分),作者使用这四种产品分别构建了一个示例业务应用程序。作者使用的示例应用程序非常小,可以由一个人在很短的时间内构建完成。本书强调了良好GUI设计的重要性,建议遵循GUI标准,鼓励使用有意义的变量名,并强调了在系统中构建在线帮助的重要性。作者相当详细地讨论了硬件和软件需求;这对那些考虑购买书中提到的产品的人很有用。在描述SQLWindows时,作者简要地提到了项目管理,但只是与产品中可用的设施有关。作者鼓励交互式开发,但不幸的是,他们通过原型和对应用程序进行修改来进行开发。作者几乎没有花时间讨论需要项目团队的大型系统的开发。第一次看到这本书时,我认为它包含了CS/OO/GUI环境的良好设计策略——这是一个重要的考虑因素。偶尔,作者会讨论带有可重用函数和对象的共享库。同样,这也不是任何真正的设计策略的一部分。和nowtesting。不幸的是,书又掉了下来。没有提到测试策略和测试数据。事实上,这本书混淆了测试和调试,并继续描述了产品中允许用户/程序员调试应用程序的所有可用工具。作者试图用知名的4GL集成开发环境产品“来说明开发一个大规模的CS应用实例”。事实上,他们煞费苦心地详细描述了一个非常小的示例系统的构造——包括每一次鼠标点击和双击。这一过程的繁琐细节占据了全书70%的篇幅。使用4GL工作台来说明原型和实现阶段的一小部分的更通用的方法将是更可取的。这本书在许多方面都是独立的,对于使用作者所讨论的四种产品之一的任何人来说,可以作为介绍性参考文本。如果能在书的最后看到一个总结就好了——但不幸的是,事实并非如此。考虑到作者写这本书的路径,我希望对所使用的四种产品进行深入的比较和回顾;相反,大多数页面只是简单地描述按什么按钮、何时按以及为什么按。看来这本书是匆匆拼凑而成的。我期望从这本书中获得的一件事是没有的-基于gui的设计和CS应用程序开发的良好结构化方法。
{"title":"UI-based design and development for client/server applications [Book Reviews]","authors":"M. Trayton","doi":"10.1109/M-PDT.1995.414849","DOIUrl":"https://doi.org/10.1109/M-PDT.1995.414849","url":null,"abstract":"In this text the authors state that they use the top 4GL integrated development environment products “to illustrate the development of a large-scale client/server application example.” Beginning very enthusiastically, the authors talk about the book as if it is the outcome of a very successful project. The authors intend the audience to be mainframe professionals who need help translating their skills for client-server, object-oriented, and graphical user interface applications (I abbreviate these terms as CS/OO/GUI). The book starts by using abbreviations that only computer professionals with a mainframe background would recognize. As the book moves on, CS/OO/GUI jargon becomes commonplace. The authors should have included a glossary. As it is, the reader must continually review previous pages to find the meaning of acronyms and abbreviations. This is annoying and it becomes clear that the reader needs quite a lot of knowledge of CS/OO/GUI to understand what the authors are saying. The authors make the assumption that, for the foreseeable future, CS/oO/GUI is the way to go in data processing and that the important new skills involve the “new wave” of CS/OO/GUI 4GL workbenches, such as PowerBuilder, Visual Basic, SQLWindows, and PARTS Workbench. The book includes discussions on the pros and cons of corporate mainframes, centralized data processing, multiple parallel processor machines, midrange minicomputers, workstations, and personal computers (“The key to CS”). The text then goes on to explain the different operating systems related to each type of hardware platform and their possibilities for the future. The authors consider relational databases to be the “cornerstone” of CS computing. They suggest that an insight into computer communication is critical in understanding CS. There then follows a general overview of computer communications and the background to CS systems. After an overview of CS/OO/GUI (this can be skipped by the more enlightened reader), the authors build a sample business application with each of the four products. The example application that the authors use is small enough to be built in a short amount of time by one person. The book emphasizes the importance of good GUI design, recommends adhering to GUI standards, encourages the use of meaningful variable names, and stresses the importance of building on-line help into a system. The authors discuss the hardware and software requirements in considerable detail; this is useful for those thinking about buying one of the products mentioned in the book. When describing SQLWindows, the authors briefly mention project management, but only in relation to the facilities available for this in the product. The authors encourage interactive development, but unfortunately they approach development by prototyping and hacking the application together. The authors spend almost no time discussing the development of larger systems that require a project team. On first seeing the book, I t","PeriodicalId":325213,"journal":{"name":"IEEE Parallel & Distributed Technology: Systems & Applications","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115259449","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Architectures with parallel I/O subsystems 具有并行I/O子系统的体系结构
Pub Date : 1900-01-01 DOI: 10.1109/m-pdt.1995.414860
Here are some examples, in approximate chronological order, of massively parallel machines that include a parallel I/O subsystem: 0 Intel iPSC hypercubes: Each hyper-cube node has an extra link that allows an YO processor to hook onto it. Thus, the number of 1/0 processors can grow to the number of hypercube nodes. In the latest version (the iPSC/860), hypercube nodes are based on the i860 microprocessor , whereas 1/0 processors use an 803 86 chip. Each I/O processor has a SCSI bus with one or more disks, and services requests from all hypercube nodes. Requests and data are routed through the node to which the 1/0 processor connects. 0 nCube hypercubes: Like the iPSC, nodes have an e m connection to an YO processor. Each VO processor connects directly to up to eight nodes.' The processors use a proprietary chip design. MasPar: A SIMD machine with up to 16K processors.' A grid and a three-stage router network connect the processors. The router also connects to a special IORAM of up to 1 Gbyte. This allows permutation of the data between the processor array and the I O W. The I O W , in turn, connects to multiple disk arrays via an YO channel. Each disk array is a RAID 3 arrangement with eight data disks and one parity disk. Intel Paragon XP/S: A mesh-suuc-tured machine that allows different configurations of compute nodes and U 0 nodes. Compute nodes are based on the 8 6 0 microprocessor. Typically, the VO nodes are concentrated in one or more rectangular I/O partitions. The Paragon is based on experience with the Touchstone Delta prototype, a 16 x 36 mesh with 5 13 processing nodes and 42 VO nodes (32 with disks and 10 with tape^).^ kSR1: A multiprocessor based on the Allcache memory design, with up to 1,088 custom processors. Each processor can connect to an adapter for external communications. One of the options is the Multiple Channel Disk adapter, which has five SCSI controllers. Each node can have up to 20 disks attached to it, in increments of five. Software configuration allows nodes with VO devices to be used exclusively for VO, or also for computation. Thinking Machines CM-Y: A multi-computer based on a fat-tree network and Sparc nodes with optional vector units. I/O is provided by a scalable disk array, which is implemented as a separate partition of disk-storage nodes4 Each …
以下是一些包含并行I/O子系统的大规模并行机器的示例,按大致时间顺序排列:0个Intel iPSC超立方体:每个超立方体节点都有一个额外的链接,允许YO处理器连接到它。因此,1/0处理器的数量可以增长到超立方体节点的数量。在最新版本(iPSC/860)中,超立方体节点基于i860微处理器,而1/0处理器使用80386芯片。每个I/O处理器都有一个带有一个或多个磁盘的SCSI总线,并处理来自所有超立方体节点的请求。请求和数据通过1/0处理器连接的节点路由。nCube超立方体:与iPSC一样,节点具有到YO处理器的e - m连接。每个VO处理器直接连接多达8个节点。”处理器使用专有的芯片设计。MasPar: SIMD机器,最多16K个处理器。网格和三级路由器网络连接处理器。路由器还连接到一个高达1gb的特殊IORAM。这允许在处理器阵列和io W之间进行数据排列,io W又通过YO通道连接到多个磁盘阵列。每个磁盘阵列都是由8个数据磁盘和1个校验磁盘组成的RAID 3。Intel Paragon XP/S:一种网格结构的机器,允许不同配置的计算节点和u0节点。计算节点基于860微处理器。通常,VO节点集中在一个或多个矩形I/O分区中。Paragon是基于Touchstone Delta原型的经验,一个16 x 36网格,有5个13个处理节点和42个VO节点(32个带磁盘,10个带磁带^)。^ kSR1:基于全缓存存储器设计的多处理器,多达1,088个自定义处理器。每个处理器可以连接到一个适配器进行外部通信。其中一个选项是多通道磁盘适配器,它有五个SCSI控制器。每个节点最多可以连接20个磁盘,以5为增量。软件配置允许带有VO设备的节点专门用于VO,或者也用于计算。Thinking Machines CM-Y:基于胖树网络和Sparc节点的多台计算机,带有可选的矢量单元。I/O由可扩展的磁盘阵列提供,该磁盘阵列作为磁盘存储节点的单独分区实现。
{"title":"Architectures with parallel I/O subsystems","authors":"","doi":"10.1109/m-pdt.1995.414860","DOIUrl":"https://doi.org/10.1109/m-pdt.1995.414860","url":null,"abstract":"Here are some examples, in approximate chronological order, of massively parallel machines that include a parallel I/O subsystem: 0 Intel iPSC hypercubes: Each hyper-cube node has an extra link that allows an YO processor to hook onto it. Thus, the number of 1/0 processors can grow to the number of hypercube nodes. In the latest version (the iPSC/860), hypercube nodes are based on the i860 microprocessor , whereas 1/0 processors use an 803 86 chip. Each I/O processor has a SCSI bus with one or more disks, and services requests from all hypercube nodes. Requests and data are routed through the node to which the 1/0 processor connects. 0 nCube hypercubes: Like the iPSC, nodes have an e m connection to an YO processor. Each VO processor connects directly to up to eight nodes.' The processors use a proprietary chip design. MasPar: A SIMD machine with up to 16K processors.' A grid and a three-stage router network connect the processors. The router also connects to a special IORAM of up to 1 Gbyte. This allows permutation of the data between the processor array and the I O W. The I O W , in turn, connects to multiple disk arrays via an YO channel. Each disk array is a RAID 3 arrangement with eight data disks and one parity disk. Intel Paragon XP/S: A mesh-suuc-tured machine that allows different configurations of compute nodes and U 0 nodes. Compute nodes are based on the 8 6 0 microprocessor. Typically, the VO nodes are concentrated in one or more rectangular I/O partitions. The Paragon is based on experience with the Touchstone Delta prototype, a 16 x 36 mesh with 5 13 processing nodes and 42 VO nodes (32 with disks and 10 with tape^).^ kSR1: A multiprocessor based on the Allcache memory design, with up to 1,088 custom processors. Each processor can connect to an adapter for external communications. One of the options is the Multiple Channel Disk adapter, which has five SCSI controllers. Each node can have up to 20 disks attached to it, in increments of five. Software configuration allows nodes with VO devices to be used exclusively for VO, or also for computation. Thinking Machines CM-Y: A multi-computer based on a fat-tree network and Sparc nodes with optional vector units. I/O is provided by a scalable disk array, which is implemented as a separate partition of disk-storage nodes4 Each …","PeriodicalId":325213,"journal":{"name":"IEEE Parallel & Distributed Technology: Systems & Applications","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125416140","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Taraflops into laptops Taraflops进入笔记本电脑
Pub Date : 1900-01-01 DOI: 10.1109/M-PDT.1994.329787
S. Wallach
BANDWIDTH W e need a t least 100 Mbyte/sec/node, which after the normal expansion for head-ers and ECC is around 1 Gbidsec of raw data on the link. This represents 22 T3 (44.736-Mbidsec) interfaces per node! LATENCY W e need an end-to-end latency through the switch network which is in line with the rest of the store hierarchy. If we look a t current processors, we see performance characteristics something like this for the different levels of the store hierarchy: Level Clocks Slowdown Register 1 Level 1 cache 2-3 2-3 Level 2 cache 6-10 2-3 Store 2 0+ 2-3 So each level down the hierarchy is a factor of 2 or 3 slower than the previous one. If we view store accessed over the switch as the next level of the memory hierarchy, this implies that we want to achieve an access through the switch in around 40-60 CPU cycles-that is, in 400-600 nanoseconds for a 1 00-MHz clocked C P U (probably a low estimate). ATiM is currently viewed as the lowest latency nonproprietary switch structure, but such switches have a single switch latency of around 1.25 sec; this implies a full switch network latency of around 4 Fsec for a 256-node machine, a factor of 10 too large. So far I have ignored the latency in getting from a user request out to the switch network. If the network is accessed as a communications device (as will happen with a naive ATM interface), this will involve system calls and the kernel of the operating system. Many thousands of instructions will be executed, translating Teraflops into laptops Stl?UP WUllUCh. COYlZE'X At a recent meeting of the High Performance Computing and Communications and Information Technology Subconi-mittee, the topic was software for scalable parallel processing. Various suppliers of hardware systems and software applications participated, including me. The consensus was that standard third-party software was beginning to emerge on scalable parallel processors, and that as a result, a new world of computing was coming. One participant went so far as to state that " one day we will run parallelized finite element code on a laptop. " I share the same view: Scalable parallel processing (SPP) will be the norm, and will pervade all computing from the laptop to the teraflop. For server systems costing $50,000 or more, parallel processors will be standard in the next year, with price erosion of 1 …
带宽我们需要至少100 Mbyte/sec/节点,在头和ECC正常扩展后,链路上的原始数据约为1 gbbidsec。这表示每个节点有22个T3 (44.736-Mbidsec)接口!我们需要通过交换机网络的端到端延迟,这与存储层次结构的其余部分一致。如果我们看一下当前的处理器,我们会看到不同级别存储层次结构的性能特征如下:级别时钟减慢寄存器1 1级缓存2-3 - 2- 2级缓存6-10 - 2-3存储2 0+ 2-3所以每个级别都比前一个慢2或3倍。如果我们将通过交换机访问的存储视为内存层次结构的下一层,这意味着我们希望通过交换机在大约40-60个CPU周期内实现访问——也就是说,对于一个以100 - mhz为时钟的CPU,在400-600纳秒内(可能是一个低估计)。ATiM目前被认为是延迟最低的非专有交换机结构,但这种交换机的单个交换机延迟约为1.25秒;这意味着对于一台256个节点的机器来说,完整的交换网络延迟大约是4秒,这是延迟的10倍。到目前为止,我忽略了从用户请求到交换机网络的延迟。如果网络是作为通信设备访问的(就像简单的ATM接口一样),这将涉及系统调用和操作系统内核。成千上万的指令将被执行,将万亿次浮点运算转换到笔记本电脑上。WUllUCh。在高性能计算、通信和信息技术小组委员会最近的一次会议上,主题是用于可扩展并行处理的软件。各种硬件系统和软件应用的供应商参与其中,包括我。大家一致认为,标准的第三方软件开始出现在可伸缩的并行处理器上,因此,一个新的计算世界即将到来。一位与会者甚至表示,“有一天我们将在笔记本电脑上运行并行有限元代码。”我也有同样的看法:可扩展并行处理(SPP)将成为常态,并将渗透到从笔记本电脑到万亿次浮点运算的所有计算中。对于成本在5万美元或以上的服务器系统,并行处理器将在明年成为标准配置,价格将下降1%。
{"title":"Taraflops into laptops","authors":"S. Wallach","doi":"10.1109/M-PDT.1994.329787","DOIUrl":"https://doi.org/10.1109/M-PDT.1994.329787","url":null,"abstract":"BANDWIDTH W e need a t least 100 Mbyte/sec/node, which after the normal expansion for head-ers and ECC is around 1 Gbidsec of raw data on the link. This represents 22 T3 (44.736-Mbidsec) interfaces per node! LATENCY W e need an end-to-end latency through the switch network which is in line with the rest of the store hierarchy. If we look a t current processors, we see performance characteristics something like this for the different levels of the store hierarchy: Level Clocks Slowdown Register 1 Level 1 cache 2-3 2-3 Level 2 cache 6-10 2-3 Store 2 0+ 2-3 So each level down the hierarchy is a factor of 2 or 3 slower than the previous one. If we view store accessed over the switch as the next level of the memory hierarchy, this implies that we want to achieve an access through the switch in around 40-60 CPU cycles-that is, in 400-600 nanoseconds for a 1 00-MHz clocked C P U (probably a low estimate). ATiM is currently viewed as the lowest latency nonproprietary switch structure, but such switches have a single switch latency of around 1.25 sec; this implies a full switch network latency of around 4 Fsec for a 256-node machine, a factor of 10 too large. So far I have ignored the latency in getting from a user request out to the switch network. If the network is accessed as a communications device (as will happen with a naive ATM interface), this will involve system calls and the kernel of the operating system. Many thousands of instructions will be executed, translating Teraflops into laptops Stl?UP WUllUCh. COYlZE'X At a recent meeting of the High Performance Computing and Communications and Information Technology Subconi-mittee, the topic was software for scalable parallel processing. Various suppliers of hardware systems and software applications participated, including me. The consensus was that standard third-party software was beginning to emerge on scalable parallel processors, and that as a result, a new world of computing was coming. One participant went so far as to state that \" one day we will run parallelized finite element code on a laptop. \" I share the same view: Scalable parallel processing (SPP) will be the norm, and will pervade all computing from the laptop to the teraflop. For server systems costing $50,000 or more, parallel processors will be standard in the next year, with price erosion of 1 …","PeriodicalId":325213,"journal":{"name":"IEEE Parallel & Distributed Technology: Systems & Applications","volume":"405 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134474325","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Parallel & Distributed Technology: Systems & Applications
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1