Muutke küpsiste eelistusi

E-raamat: Tools and Environments for Parallel and Distributed Computing

Edited by , Edited by
Teised raamatud teemal:
  • Formaat - PDF+DRM
  • Hind: 125,91 €*
  • * hind on lõplik, st. muud allahindlused enam ei rakendu
  • Lisa ostukorvi
  • Lisa soovinimekirja
  • See e-raamat on mõeldud ainult isiklikuks kasutamiseks. E-raamatuid ei saa tagastada.
  • Raamatukogudele
Teised raamatud teemal:

DRM piirangud

  • Kopeerimine (copy/paste):

    ei ole lubatud

  • Printimine:

    ei ole lubatud

  • Kasutamine:

    Digitaalõiguste kaitse (DRM)
    Kirjastus on väljastanud selle e-raamatu krüpteeritud kujul, mis tähendab, et selle lugemiseks peate installeerima spetsiaalse tarkvara. Samuti peate looma endale  Adobe ID Rohkem infot siin. E-raamatut saab lugeda 1 kasutaja ning alla laadida kuni 6'de seadmesse (kõik autoriseeritud sama Adobe ID-ga).

    Vajalik tarkvara
    Mobiilsetes seadmetes (telefon või tahvelarvuti) lugemiseks peate installeerima selle tasuta rakenduse: PocketBook Reader (iOS / Android)

    PC või Mac seadmes lugemiseks peate installima Adobe Digital Editionsi (Seeon tasuta rakendus spetsiaalselt e-raamatute lugemiseks. Seda ei tohi segamini ajada Adober Reader'iga, mis tõenäoliselt on juba teie arvutisse installeeritud )

    Seda e-raamatut ei saa lugeda Amazon Kindle's. 

Surveys current software tools for developing, implementing, and supporting high-performance parallel and distributed computing environments and applications. Four of the six contributions describe message-passing models for network-centric applications, distributed shared memory tools, three specific distributed-object approaches, and several grid architectures. Ping, concurrency, and numerical computation experiments compare the performance of the Java-RMI, CORBA, and DCOM distributed-object technologies. The last chapter presents a software development process for high-performance parallel and distributed computing environments and a sample parallel model of stock option pricing. Annotation ©2004 Book News, Inc., Portland, OR (booknews.com)

  • An invaluable reference for anyone designing new parallel or distributed systems.
  • Includes detailed case studies of specific systems from Stanford, MIT, and other leading research universities.
  • The authors emphasize performance, surveying all available techniques.

Arvustused

"...an excellent reference for tools and paradigms for building high performance parallel and distributed computing systems..." (Computing Reviews.com, February 3, 2005) "This book is recommended for libraries supporting graduate level programs in computer science and programs utilizing high-performance computing in their research." (E-STREAMS, August 2004)

Preface.
1. Parallel and Distributed Computing (S. Hariri & M. Parashar).
1.1 Introduction: Basic Concepts.
1.2 Promises and Challenges of Parallel and Distributed Systems.
1.2.1 Processing Technology.
1.2.2 Networking Technology.
1.2.3 Software Tools and Environments.
1.3 Distributed System Design Framework.
References and Further Reading.
2. Message-Passing Tools (S. Hariri & I. Ra).
2.1 Introduction.
2.2 Message-Passing Tools versus Distributed Shared Memory.
2.2.1 Distributed Shared Memory Model.
2.2.2 Message-Passing Model.
2.3 Message-Passing System: Desirable Features.
2.4 Classification of Message-Passing Tools.
2.4.1 Classification by Implementation.
2.5 Overview of Message-Passing Tools.
2.5.1 Socket-Based Message Passing.
2.5.2 p4.
2.5.3 Parallel Virtual Machine.
2.5.4 Message-Passing Interface.
2.5.5 Nexus.
2.5.6 Madeleine I and II.
2.5.7 Active Messages.
2.6 ACS.
2.6.1 Multithread Communications Services.
2.6.2 Separation of Data and Control Functions.
2.6.3 Programmable Communication, Control, and Management Service.
2.6.4 Multiple Communication Interfaces.
2.6.5 Adaptive Group Communication Services.
2.7 Experimental Results and Analysis.
2.7.1 Experimental Environment.
2.7.2 Performance of Primitives.
2.7.3 Application Performance Benchmarking.
2.7.4 Performance Results of Adaptive Schemes.
2.8 Conclusions.
References.
3. Distributed Shared Memory Tools (M. Parashar & S. Chandra).
3.1 Introduction.
3.2 Cache Coherence.
3.2.1 Directory-Based Cache Coherence.
3.3 Shared Memory Consistency Models.
3.4 Distributed Memory Architectures.
3.5 Classification of Distributed Shared Memory Systems.
3.5.1 Hardware-Based DSM Systems.
3.5.2 Mostly Software Page-Based DSM Systems.
3.5.3 All-Software Object-Based DSM Systems.
References.
4. Distributed-Object Computing Tools (R. Raje, et al.).
4.1 Introduction.
4.2 Basic Model.
4.2.1 RMI.
4.2.2 CORBA.
4.2.3 DCOM.
4.3 Examples.
4.3.1 Experimental Setup.
4.3.2 Developing Applications under RMI, CORBA, and DCOM.
4.3.3 Experiment 1: Ping.
4.3.4 Experiment 2: Producer-Consumer Problem.
4.3.5 Experiment 3: Numerical Computation.
4.4 Comparison of the Three Paradigms.
4.4.1 Dependency Issues.
4.4.2 Implementation Details.
4.4.3 Architecture Details.
4.4.4 Support for Additional Features.
4.4.5 Performance Comparison.
4.5 Conclusions.
References.
5. Gestalt of the Grid (G. von Laszewski & P. Wagstrom).
5.1 Introduction.
5.1.1 Motivation.
5.1.2 Enabling Factors.
5.2 Definitions.
5.3 Multifaceted Grid Architecture.
5.3.1 N-Tiered Grid Architecture.
5.3.2 Role-Based Grid Architecture.
5.3.3 Service-Based Grid Architecture.
5.3.4 Grid Challenges.
5.4 Grid Management Aspects.
5.4.1 Managing Grid Security.
5.4.2 Managing Grid Information.
5.4.3 Managing Grid Data.
5.4.4 Managing Grid Execution and Resources.
5.4.5 Managing Grid Software.
5.4.6 Managing Grid Hardware.
5.5 Grid Activities.
5.5.1 Community Activities.
5.5.2 Grid Middleware.
5.5.3 High-Throughput Computing.
5.6 Grid Applications.
5.6.1 Astrophysics Simulation Collaboratory.
5.6.2 Particle Physics Data Grid.
5.6.3 NEESgrid.
5.7 Portals.
5.7.1 HotPage.
5.7.2 Webflow and Gateway.
5.7.3 XCAT.
5.7.4 UNICORE.
5.7.5 JiPANG.
5.7.6 PUNCH.
5.7.7 Access Grid.
5.7.8 Commercial Grid Activities.
5.8 Conclusions.
References.
6. Software Development for Parallel and Distributed Computing (M. Parashar & S. Hariri).
6.1 Introduction.
6.2 Issues in HPC Software Development.
6.2.1 Models for Parallel Computation.
6.2.2 Portable Application Description Medium.
6.2.3 Parallel Algorithm Development.
6.2.4 Program Implementation and Runtime.
6.2.5 Visualization and Animation.
6.2.6 Maintainability.
6.2.7 Reliability.
6.2.8 Reusability.
6.3 HPC Software Development Process.
6.4 Parallel Modeling of Stock Option Pricing.
6.5 Inputs.
6.6 Application Analysis Stage.
6.7 Application Development Stage.
6.7.1 Algorithm Development Module.
6.7.2 System-Level Mapping Module.
6.7.3 Machine-Level Mapping Module.
6.7.4 Implementation/Coding Module.
6.7.5 Design Evaluator Module.
6.8 Compile-Time and Runtime Stage.
6.9 Evaluation Stage.
6.10 Maintenance/Evolution Stage.
6.11 Existing Software Support.
6.11.1 Application Specifications Filter.
6.11.2 Application Analysis Stage.
6.11.3 Application Development Stage.
6.11.4 Compile-Time and Runtime Stage.
6.11.5 Evaluation Stage.
6.11.6 Maintenance/Evolution Stage.
References. 
Index.


SALIM HARIRI, PhD, is a professor in the Department of Electrical and Computer Engineering at the University of Arizona, Tucson, where he directs the High Performance Distributed Computing Laboratory. He is the Editor in Chief for the Cluster Computing Journal and the founder of IEEE International Symposium on High Performance Distributed Computing (HPDC). He is coauthor/editor of three books on parallel and distributed computing and has published over a hundred journal articles and conference papers. MANISH PARASHAR, PhD, is an associate professor in the Department of Electrical and Computer Engineering at Rutgers, The State University of New Jersey, where he is Director of The Applied Software Systems Laboratory (TASSL). Professor Parashar is a recipient of the NSF CAREER award and the Enrico Fermi scholarship, and is a senior member of the IEEE. In addition to publishing over a hundred technical papers in international journals and conferences, Professor Parashar has coauthored/edited three books, and has contributed to several others, in the area of parallel and distributed computing.