Sciweavers

Share
CLUSTER
2006
IEEE

Optimizing MPI collective communication by orthogonal structures

9 years 1 months ago
Optimizing MPI collective communication by orthogonal structures
Many parallel applications from scientific computing use MPI collective communication operations to collect or distribute data. Since the execution times of these communication operations increase with the number of participating processors, scalability problems might occur. In this article, we show for different MPI implementations how the execution time of collective communication operations can be significantly improved by a restructuring based on orthogonal processor structures with two or more levels. As platform, we consider a dual Xeon cluster, a Beowulf cluster and a Cray T3E with different MPI implementations. We show that the execution time of operations like MPI Bcast or MPI Allgather can be reduced by 40% and 70% on the dual Xeon cluster and the Beowulf cluster. But also on a Cray T3E a significant improvement can be obtained by a careful selection of the processor groups. We demonstrate that the optimized communication operations can be used to reduce the execution time o...
Matthias Kühnemann, Thomas Rauber, Gudula R&u
Added 11 Dec 2010
Updated 11 Dec 2010
Type Journal
Year 2006
Where CLUSTER
Authors Matthias Kühnemann, Thomas Rauber, Gudula Rünger
Comments (0)
books