Sparse collective operations for MPI

Abstract

We discuss issues in designing sparse (nearest neighbor) collective operations for communication and reduction operations in small neighborhoods for the Message Passing Interface (MPI).We propose three such operations, namely a sparse gather operation, a sparse all-to-all, and a sparse reduction operation in both regular and irregular (vector) variants. By two simple experiments we show a) that a collective handle for message scheduling and communication optimization is necessary for any such interface, b) that the possibly different amount of communication between neighbors need to be taken into account by the optimization, and c) illustrate the improvements that are possible by schedules that posses global information compared to implementations that can rely on only local information. We discuss different forms the interface and optimization handles could take. The paper is inspired by current discussion in the MPI Forum.

DOI: 10.1109/IPDPS.2009.5160935

Extracted Key Phrases

9 Figures and Tables

Cite this paper

@article{Hoefler2009SparseCO, title={Sparse collective operations for MPI}, author={Torsten Hoefler and Jesper Larsson Tr{\"a}ff}, journal={2009 IEEE International Symposium on Parallel & Distributed Processing}, year={2009}, pages={1-8} }