Tools and Environments for Parallel and Distributed Computing
Wiley Series on Parallel and Distributed Computing, Band 34 1. Aufl.
An invaluable reference for anyone designing new parallel or distributed systems. Includes detailed case studies of specific systems from Stanford, MIT, and other leading research universities. The authors emphasize performance, surveying all available techniques.
Preface. 1. Parallel and Distributed Computing (S. Hariri & M. Parashar). 1.1 Introduction: Basic Concepts. 1.2 Promises and Challenges of Parallel and Distributed Systems. 1.2.1 Processing Technology. 1.2.2 Networking Technology. 1.2.3 Software Tools and Environments. 1.3 Distributed System Design Framework. References and Further Reading. 2. Message-Passing Tools (S. Hariri & I. Ra). 2.1 Introduction. 2.2 Message-Passing Tools versus Distributed Shared Memory. 2.2.1 Distributed Shared Memory Model. 2.2.2 Message-Passing Model. 2.3 Message-Passing System: Desirable Features. 2.4 Classification of Message-Passing Tools. 2.4.1 Classification by Implementation. 2.5 Overview of Message-Passing Tools. 2.5.1 Socket-Based Message Passing. 2.5.2 p4. 2.5.3 Parallel Virtual Machine. 2.5.4 Message-Passing Interface. 2.5.5 Nexus. 2.5.6 Madeleine I and II. 2.5.7 Active Messages. 2.6 ACS. 2.6.1 Multithread Communications Services. 2.6.2 Separation of Data and Control Functions. 2.6.3 Programmable Communication, Control, and Management Service. 2.6.4 Multiple Communication Interfaces. 2.6.5 Adaptive Group Communication Services. 2.7 Experimental Results and Analysis. 2.7.1 Experimental Environment. 2.7.2 Performance of Primitives. 2.7.3 Application Performance Benchmarking. 2.7.4 Performance Results of Adaptive Schemes. 2.8 Conclusions. References. 3. Distributed Shared Memory Tools (M. Parashar & S. Chandra). 3.1 Introduction. 3.2 Cache Coherence. 3.2.1 Directory-Based Cache Coherence. 3.3 Shared Memory Consistency Models. 3.4 Distributed Memory Architectures. 3.5 Classification of Distributed Shared Memory Systems. 3.5.1 Hardware-Based DSM Systems. 3.5.2 Mostly Software Page-Based DSM Systems. 3.5.3 All-Software Object-Based DSM Systems. References. 4. Distributed-Object Computing Tools (R. Raje, et al.). 4.1 Introduction. 4.2 Basic Model. 4.2.1 RMI. 4.2.2 CORBA. 4.2.3 DCOM. 4.3 Examples. 4.3.1 Experimental Setup. 4.3.2 Developing Applications under RMI, CORBA, and DCOM. 4.3.3 Experiment 1: Ping. 4.3.4 Experiment 2: Producer-Consumer Problem. 4.3.5 Experiment 3: Numerical Computation. 4.4 Comparison of the Three Paradigms. 4.4.1 Dependency Issues. 4.4.2 Implementation Details. 4.4.3 Architecture Details. 4.4.4 Support for Additional Features. 4.4.5 Performance Comparison. 4.5 Conclusions. References. 5. Gestalt of the Grid (G. von Laszewski & P. Wagstrom). 5.1 Introduction. 5.1.1 Motivation. 5.1.2 Enabling Factors. 5.2 Definitions. 5.3 Multifaceted Grid Architecture. 5.3.1 N-Tiered Grid Architecture. 5.3.2 Role-Based Grid Architecture. 5.3.3 Service-Based Grid Architecture. 5.3.4 Grid Challenges. 5.4 Grid Management Aspects. 5.4.1 Managing Grid Security. 5.4.2 Managing Grid Information. 5.4.3 Managing Grid Data. 5.4.4 Managing Grid Execution and Resources. 5.4.5 Managing Grid Software. 5.4.6 Managing Grid Hardware. 5.5 Grid Activities. 5.5.1 Community Activities. 5.5.2 Grid Middleware. 5.5.3 High-Throughput Computing. 5.6 Grid Applications. 5.6.1 Astrophysics Simulation Collaboratory. 5.6.2 Particle Physics Data Grid. 5.6.3 NEESgrid. 5.7 Portals. 5.7.1 HotPage. 5.7.2 Webflow and Gateway. 5.7.3 XCAT. 5.7.4 UNICORE. 5.7.5 JiPANG. 5.7.6 PUNCH. 5.7.7 Access Grid. 5.7.8 Commercial Grid Activities. 5.8 Conclusions. References. 6. Software Development for Parallel and Distributed Computing (M. Parashar & S. Hariri). 6.1 Introduction. 6.2 Issues in HPC Software Development. 6.2.1 Models for Parallel Computation. 6.2.2 Portable Application Description Medium. 6.2.3 Parallel Algorithm Development. 6.2.4 Program Implementation and Runtime. 6.2.5 Visualization and Animation. 6.2.6 Maintainability. 6.2.7 Reliability. 6.2.8 Reusability. 6.3 HPC Software Development Process. 6.4 Parallel Modeling of Stock Option Pricing. 6.5 Inputs. 6.6 Application Analysis Stage. 6.7 Application Development Stage. 6.7.1 Algorithm Development Module. 6.7.2 System-Level Mapping Module. 6.7.3 Machine-Level Mapping Module. 6.7.4 Implementation/Coding Module. 6.7.5 Design Evaluator Module. 6.8 Compile-Time and Runtime Stage. 6.9 Evaluation Stage. 6.10 Maintenance/Evolution Stage. 6.11 Existing Software Support. 6.11.1 Application Specifications Filter. 6.11.2 Application Analysis Stage. 6.11.3 Application Development Stage. 6.11.4 Compile-Time and Runtime Stage. 6.11.5 Evaluation Stage. 6.11.6 Maintenance/Evolution Stage. References. Index.
"…an excellent reference for tools and paradigms for building high performance parallel and distributed computing systems…" (Computing Reviews.com, February 3, 2005) "This book is recommended for libraries supporting graduate level programs in computer science and programs utilizing high-performance computing in their research.” (E-STREAMS, August 2004)
SALIM HARIRI, PhD, is a professor in the Department of Electrical and Computer Engineering at the University of Arizona, Tucson, where he directs the High Performance Distributed Computing Laboratory. He is the Editor in Chief for the Cluster Computing Journal and the founder of IEEE International Symposium on High Performance Distributed Computing (HPDC). He is coauthor/editor of three books on parallel and distributed computing and has published over a hundred journal articles and conference papers. MANISH PARASHAR, PhD, is an associate professor in the Department of Electrical and Computer Engineering at Rutgers, The State University of New Jersey, where he is Director of The Applied Software Systems Laboratory (TASSL). Professor Parashar is a recipient of the NSF CAREER award and the Enrico Fermi scholarship, and is a senior member of the IEEE. In addition to publishing over a hundred technical papers in international journals and conferences, Professor Parashar has coauthored/edited three books, and has contributed to several others, in the area of parallel and distributed computing.
Today's most complete reference on the rapidly evolving software technology used in parallel and distributed computing Designing, developing, and utilizing the kinds of versatile computing environments that will effectively support today's complex scientific and engineering applications is an ongoing research challenge. Tools and Environments for Parallel and Distributed Computing is the first comprehensive survey of the successful computing methodologies used in today's scientific communities, including both a summary of past techniques and an introduction to future technologies. Covering a wide variety of approaches to parallel and distributed computing, with emphasis on such factors as tradeoffs between performance and cost, the authors present a thorough survey of today's most promising software applications, their strengths and limitations, including: Grid computing Message-passing tools Distributed shared-memory tools Distributed object computing tools Each methodology is supported by detailed case studies. An important hands-on reference for researchers, educators, and practitioners in the field, this is a practical guide for the development and selection of today's scientific and engineering software systems.
Diese Produkte könnten Sie auch interessieren:
NeuheitenA Companion to Literary Theory 35,99 €
Nursing Older People at a Glance 26,99 €
Annual Update for Accountants and A... 77,99 €
Statement of Cash Flows: Preparatio... 77,99 €
Electrochemical Engineering 108,99 €