Performance models for MPI collective communications with network contention

Hyacinthe Nzigou Mamadou, Takeshi Nanri, Kazuaki Murakami

    Research output: Contribution to journalArticlepeer-review

    Abstract

    The paper presents a novel approach to estimate the performance of MPI collective communications. Our objective is to help researchers to make appropriate decisions on their message-passing applications. For each collective communication, we attempt to apply LogGP and P-LogP standard point-to-point models. The resulted models are compared with the empirical data in order to identify the most suitable for performance characterization of collective operations. For the communications on large clusters with large size messages, the network contention problem can significantly affect the performance. Hence, to reduce the relative gap between the prediction and the measured runtime, the contention issue is also modeled, by a queuing theory analysis method, and taken in account with the total performance estimation. The experiments performed on a cluster which consists of 64 processors interconnected by Gigabit Ethernet network show encouraging results. For any collective operation, given a number of processors and a range of message sizes, there is at least one model that predicts the performance precisely. We could achieve a gap between the predicted and the measured run-time around 15%. Thus, by handling the contention problem, we could reduce around 80% of the relative gap.

    Original languageEnglish
    Pages (from-to)1015-1024
    Number of pages10
    JournalIEICE Transactions on Communications
    VolumeE91-B
    Issue number4
    DOIs
    Publication statusPublished - 2008

    All Science Journal Classification (ASJC) codes

    • Software
    • Computer Networks and Communications
    • Electrical and Electronic Engineering

    Fingerprint Dive into the research topics of 'Performance models for MPI collective communications with network contention'. Together they form a unique fingerprint.

    Cite this