[mvapich-discuss] mvapich2-1.6 problems with np ~ >= 1000

Johnny Devaprasad johnnydevaprasad at gmail.com
Tue Apr 5 06:01:37 EDT 2011


Hi all,

I am running a simple MPI program (only calls MPI_Get_processor_name).

This sometime works and most of the time does not...

mpirun_rsh -np 2000 -hostfile
/home/jd/working/simple/mvapich2/machinefile_large
/home/jd/working/simple/mvapich2/mvapich2_pgi
Exit code -5 signaled from node015
MPI process (rank: 315) terminated unexpectedly on node027
MPI process (rank: 214) terminated unexpectedly on node014
handle_mt_peer: fail to read...: Success
handle_mt_peer: fail to read...: Success
handle_mt_peer: fail to read...: Success
handle_mt_peer: fail to read...: Success
handle_mt_peer: fail to read...: Success
handle_mt_peer: fail to read...: Success
handle_mt_peer: fail to read...: Success
handle_mt_peer: fail to read...: Success
handle_mt_peer: fail to read...: Success
handle_mt_peer: fail to read...: Success

mpirun_rsh -np 1000 -hostfile
/home/jd/working/simple/mvapich2/machinefile_large
/home/jd/working/simple/mvapich2/mvapich2_pgi
MPI process (rank: 435) terminated unexpectedly on node044
Exit code -5 signaled from node041
MPI process (rank: 777) terminated unexpectedly on node048
handle_mt_peer: fail to read...: Success
handle_mt_peer: fail to read...: Success
handle_mt_peer: fail to read...: Success
handle_mt_peer: fail to read...: Success
handle_mt_peer: fail to read...: Success


Regards,
Johnny
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://mail.cse.ohio-state.edu/pipermail/mvapich-discuss/attachments/20110405/cbe003e9/attachment-0001.html


More information about the mvapich-discuss mailing list