Synopses & Reviews
A hands-on introduction to parallel programming based on the Message-Passing Interface (MPI) standard, the de-facto industry standard adopted by major vendors of commercial parallel systems. This textbook/tutorial, based on the C language, contains many fully-developed examples and exercises. The complete source code for the examples is available in both C and Fortran 77. Students and professionals will find that the portability of MPI, combined with a thorough grounding in parallel programming principles, will allow them to program any parallel system, from a network of workstations to a parallel supercomputer.
* Proceeds from basic blocking sends and receives to the most esoteric aspects of MPI.
* Includes extensive coverage of performance and debugging.
* Discusses a variety of approaches to the problem of basic I/O on parallel machines.
* Provides exercises and programming assignments.
Review
"...the detailed discussion of many complex and confusing issues makes the book an important information source for programmers developing large applications using MPI."
-—L.M. Liebrock, ACM Computing Reviews
Review
"...the detailed discussion of many complex and confusing issues makes the book an important information source for programmers developing large applications using MPI."
- L.M. Liebrock, ACM Computing Reviews
Review
"...the detailed discussion of many complex and confusing issues makes the book an important information source for programmers developing large applications using MPI."
- L.M. Liebrock, ACM Computing Reviews
Review
es the book an important information source for programmers developing large applications using MPI."
- L.M. Liebrock, ACM Computing Reviews
Synopsis
A hands-on introduction to parallel programming based on the Message-Passing Interface (MPI) standard, the de-facto industry standard adopted by major vendors of commercial parallel systems. This textbook/tutorial, based on the C language, contains many fully-developed examples and exercises. The complete source code for the examples is available in both C and Fortran 77. Students and professionals will find that the portability of MPI, combined with a thorough grounding in parallel programming principles, will allow them to program any parallel system, from a network of workstations to a parallel supercomputer.
* Proceeds from basic blocking sends and receives to the most esoteric aspects of MPI.
* Includes extensive coverage of performance and debugging.
* Discusses a variety of approaches to the problem of basic I/O on parallel machines.
* Provides exercises and programming assignments.
Synopsis
>+ Discusses a variety of approaches to the problem of basic I/O on parallel machines.
+ Provides exercises and programming assignments.
Synopsis
arallel machines.
Provides exercises and programming assignments.Description
Includes bibliographical references (p. 403-406) and index.
About the Author
Peter Pacheco received a PhD in mathematics from Florida State University. After completing graduate school, he became one of the first professors in UCLA’s “Program in Computing,” which teaches basic computer science to students at the College of Letters and Sciences there. Since leaving UCLA, he has been on the faculty of the University of San Francisco. At USF Peter has served as chair of the computer science department and is currently chair of the mathematics department. His research is in parallel scientific computing. He has worked on the development of parallel software for circuit simulation, speech recognition, and the simulation of large networks of biologically accurate neurons. Peter has been teaching parallel computing at both the undergraduate and graduate levels for nearly twenty years. He is the author of Parallel Programming with MPI, published by Morgan Kaufmann Publishers.
University of San Francisco, USA
Table of Contents
Chapter 1 Introduction
Chapter 2 An Overview of Parallel Computing
Chapter 3 Greetings!
Chapter 4 An Application: Numerical Integration
Chapter 5 Collective Communication
Chapter 6 Grouping Data for Communication
Chapter 7 Communicators and Topologies
Chapter 8 Dealing with I/O
Chapter 9 Debugging Your Program
Chapter 10 Design and Coding of Parallel Programs
Chapter 11 Performance
Chapter 12 More on Performance
Chapter 13 Advanced Point-to-Point Communication
Chapter 14 Parallel Algorithms
Chapter 15 Parallel Libraries
Chapter 16 Wrapping Up
Appendix A Summary of MPI Commands
Appendix B MPI on the Internet