Sciweavers

60 search results - page 3 / 12
» Process Mapping for MPI Collective Communications
Sort
View
IPPS
2008
IEEE
14 years 19 days ago
Scaling alltoall collective on multi-core systems
MPI Alltoall is one of the most communication intense collective operation used in many parallel applications. Recently, the supercomputing arena has witnessed phenomenal growth o...
Rahul Kumar, Amith R. Mamidala, Dhabaleswar K. Pan...
ICPP
2009
IEEE
13 years 4 months ago
Cache-Efficient, Intranode, Large-Message MPI Communication with MPICH2-Nemesis
The emergence of multicore processors raises the need to efficiently transfer large amounts of data between local processes. MPICH2 is a highly portable MPI implementation whose l...
Darius Buntinas, Brice Goglin, David Goodell, Guil...
ICPADS
2010
IEEE
13 years 4 months ago
GMH: A Message Passing Toolkit for GPU Clusters
Driven by the market demand for high-definition 3D graphics, commodity graphics processing units (GPUs) have evolved into highly parallel, multi-threaded, many-core processors, whi...
Jie Chen, William A. Watson III, Weizhen Mao
IPPS
2006
IEEE
14 years 8 days ago
Topology-aware task mapping for reducing communication contention on large parallel machines
Communication latencies constitute a significant factor in the performance of parallel applications. With techniques such as wormhole routing, the variation in no-load latencies ...
T. Agarwal, Amit Sharma, A. Laxmikant, Laxmikant V...
PVM
1998
Springer
13 years 10 months ago
SKaMPI: A Detailed, Accurate MPI Benchmark
Abstract. SKaMPI is a benchmark for MPI implementations. Its purpose is the detailed analysis of the runtime of individual MPI operations and comparison of these for di erent imple...
Ralf Reussner, Peter Sanders, Lutz Prechelt, Matth...