Abstract-- The performance of collective communication operations is known to have a significant impact on the scalability of some applications. Indeed, the global, synchronous nat...
Ron Brightwell, Sue Goudy, Arun Rodrigues, Keith D...
Abstract. Most parallel systems on which MPI is used are now hierarchical: some processors are much closer to others in terms of interconnect performance. One of the most common su...
Hao Zhu, David Goodell, William Gropp, Rajeev Thak...
We describe a generic programming model to design collective communications on SMP clusters. The programming model utilizes shared memory for collective communications and overlap...
We describe the use of MPI for writing system software and tools, an area where it has not been previously applied. By “system software” we mean collections of tools used for s...
Narayan Desai, Rick Bradshaw, Andrew Lusk, Ewing L...
Application development for distributed-computing ``Grids'' can benefit from tools that variously hide or enable application-level management of critical aspects of the ...
Nicholas T. Karonis, Brian R. Toonen, Ian T. Foste...