[mvapich-discuss] MPI_Send over 2 GB fails

Matthew Koop koop at cse.ohio-state.edu
Sun Feb 22 17:01:47 EST 2009


Vittorio,

This is a known issue we have with MVAPICH2. Currently some of the
internal data structures within the library are not large enough to handle
over 2GB of data in a single send operation.

We are planning on fixing this in a future release.

Matt

On Sun, 22 Feb 2009, Vittorio wrote:

> hello!
> i'm performing some performance test of mpvapich2 on infiniband:
> the test is very simple sending fixed quantities of data from one node to
> another.
> from 1 kB to 2 GB there are no problems but as soon as i try to transfer 4GB
> and above i get
>
> Fatal error in MPI_Send: Internal MPI error!, error stack:
> MPI_Send(192): MPI_Send(buf=0x6020a0, count=536870912, MPI_UNSIGNED_LONG,
> dest=1, tag=1, MPI_COMM_WORLD) failed
> (unknown)(): Internal MPI error![cli_0]: aborting job:
> Fatal error in MPI_Send: Internal MPI error!, error stack:
> MPI_Send(192): MPI_Send(buf=0x6020a0, count=536870912, MPI_UNSIGNED_LONG,
> dest=1, tag=1, MPI_COMM_WORLD) failed
> (unknown)(): Internal MPI error!
> rank 0 in job 11  randori_45329   caused collective abort of all ranks
>   exit status of rank 0: return code 1
>
> the two machines are equal with a 64bit OS and equipped with 32 GB of ram.
> i also tried the program on a single machine, but i receive the same error
> just after the 2 GB transfer.
>
> i'm pretty sure MPI can send more than 4 GB of data so i just can't figure
> out what the problem might be.
> any help is really appreciated
> thanks a lot
> Vittorio
>



More information about the mvapich-discuss mailing list