[mvapich-discuss] program hanged using mvapich with large number
of processes
Dhabaleswar Panda
panda at cse.ohio-state.edu
Sat Jan 23 00:03:27 EST 2010
Can you try the latest nightly tarball of the bugfix branch version of
MVAPICH2 1.4 (from the following URL) and let us know whether the issue
persists.
http://mvapich.cse.ohio-state.edu/nightly/mvapich2/branches/1.4
I am also assuming that you are using the OpenFabrics-Gen2 interface of
this release. Please confirm.
Thanks,
DK
On Sat, 23 Jan 2010, Weimin Wang wrote:
> Hello, Dhabaleswar,
>
> Thank you for your response.
>
> The version of MVAPICH2 I am using is 2-1.4. I do not know the IB adapter
> type of my cluster. When running ifconfig, I get:
>
> wmwang at node73:~/meteo/mvapich2-1.4> ifconfig -a
> ib0 Link encap:UNSPEC HWaddr
> 80-00-00-48-FE-80-00-00-00-00-00-00-00-00-00-00
> inet addr:10.10.10.73 Bcast:10.255.255.255 Mask:255.0.0.0
> inet6 addr: fe80::202:c903:5:5271/64 Scope:Link
> UP BROADCAST RUNNING MULTICAST MTU:65520 Metric:1
> RX packets:15383132 errors:0 dropped:0 overruns:0 frame:0
> TX packets:12294382 errors:0 dropped:0 overruns:0 carrier:0
> collisions:0 txqueuelen:256
> RX bytes:15352949444 (14641.7 Mb) TX bytes:130554397150 (124506.3
> Mb)
> Thank you.
>
> Bests,
> Weimin
>
> On Sat, Jan 23, 2010 at 5:52 AM, Dhabaleswar Panda <panda at cse.ohio-state.edu
> > wrote:
>
> > Can you tell us the MVAPICH2 version you are using. Also, can you tell us
> > the IB adapter type used in your system.
> >
> > Thanks,
> >
> > DK
> >
> > On Fri, 22 Jan 2010, Weimin Wang wrote:
> >
> > > Hello, list,
> > >
> > > I have got a strange problem with mvapich2. For cpi example, when I run
> > it
> > > with small number of processes, it is OK:
> > >
> > > wmwang at node32:~/test> mpirun_rsh -ssh -np 2 -hostfile ./ma ./cpi
> > > Process 0 on node32
> > > Process 1 on node32
> > > pi is approximately 3.1416009869231241, Error is 0.0000083333333309
> > > wall clock time = 0.000174
> > >
> > > wmwang at node32:~/test> mpirun_rsh -ssh -np 10 -hostfile ./ma ./cpi
> > > Process 8 on node33
> > > pi is approximately 3.1416009869231249, Error is 0.0000083333333318
> > > wall clock time = 0.000127
> > > Process 1 on node32
> > > Process 3 on node32
> > > Process 0 on node32
> > > Process 4 on node32
> > > Process 2 on node32
> > > Process 6 on node32
> > > Process 5 on node32
> > > Process 7 on node32
> > > Process 9 on node33
> > > However, when I run cpi with large number processes, the program hangs
> > with
> > > no output:
> > >
> > > wmwang at node32:~/test> mpirun_rsh -ssh -np 18 -hostfile ./ma ./cpi
> > >
> > > And top command in node32 show that,
> > >
> > > PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
> > > 14507 wmwang 15 0 60336 50m 676 S 56 0.2 0:03.86 mpispawn
> > > The system I used is,
> > >
> > > wmwang at node33:~> uname -a
> > > Linux node33 2.6.16.60-0.42.4_lustre.1.8.1.1-smp #1 SMP Fri Aug 14
> > 08:33:26
> > > MDT 2009 x86_64 x86_64 x86_64 GNU/Linux
> > > The compiler is pgi v10.0.
> > >
> > > Would you please give me any hint for this problem?
> > >
> >
> >
>
More information about the mvapich-discuss
mailing list