This document describes MadMPI benchmark installation and configuration.
For any question, mailto:
For more information, see: http://pm2.gforge.inria.fr/mpibenchmark/
A quick cheat sheet for the impatient:
./configure make mpiexec -n 2 -host host1,host2 ./mpi_bench_overlap | tee out.dat
It runs from 10 minutes through 2h, depanding on network speed. Then build the performance report using:
It outputs data in
out.dat.d/. It is possible to transfer data to another host and extract the performance report with another installation of MadMPI benchmark so as to not have to install
gnuplot on the computing nodes.
Please send the
out.dat to Alexa.email@example.com@m..Deni.firstname.lastname@example.org@email@example.com@m.r to have it integrated on the MadMPI benchmark web site.
MadMPI benchmark follows usual autoconf procedure:
./configure [your options here] make make install
make install step is optional. The benchmark may be run from its build directory.
mpi_bench_overlapon 2 nodes, capture its standard output in a file, and pass this file to
mpi_bench_extract. The processed data is outputed to a
mpi_bench_sendrecv: send/receive pingpong, used as a reference
mpi_bench_noncontig: send/receive pingpong with non-contiguous datatype, used as a reference
mpi_bench_send_overhead: processor time consumed on sender side to send data (the overhead from LogP). Usefull to explain overlap benchmarks.
mpi_bench_overlap_sender: overlap on sender side (i.e. MPI_Isend, computation, MPI_Wait), total time
mpi_bench_overlap_recv: overlap on receiver side (i.e. MPI_Irecv, computation, MPI_Wait), total time
mpi_bench_overlap_bidir: overlap on both sides
mpi_bench_overlap_sender_noncontig: overlap on sender side, with non-contiguous datatype
mpi_bench_overlap_send_overhead: overlap on sender side (i.e. MPI_Isend, computation, MPI_Wait), measure time on sender side only
mpi_bench_overlap_Nload: overlap on sender side, with multi-threaded computation load