[mvapich-discuss] process limits

Dhabaleswar Panda panda at cse.ohio-state.edu
Wed Aug 29 17:33:46 EDT 2007


Mark, 

Thanks for letting us know this issue with tcsh. We are taking a look
at this also.

Best Regards, 

DK

> DK,
>     Thanks for the quick reply.
> 
>     I'm maybe jumping ahead here, but assuming the "timeout" is the
>     one that I suggested from the 10 July patch (mentioned below)
>     I have a question related to that patch.
>     Could you/your people suggest why this patch seems to fail for
>     users running tcsh (as opposed to bash)?  My initial testing,
>     which used bash, of this patch found that it worked quite nicely
>     to clean up jobs when one or more of the processes killed
>     themselves.  An mpirun_rsh thread detected this condition and
>     then killed the remaining remote tasks.  However, we now find
>     that users employing tcsh login shell are not so lucky.  The
>     detection part of the patch works successfully and the remote
>     sshd tasks are killed successfully but the remote MPI processes
>     continue to run.  Quite interesting if it weren't so bad.  We
>     can be left with large numbers of CPU burning tasks that are
>     difficult to find and kill.
> 
>     Shell experts' ideas welcome.
> 
>             regards,
> 
> Dhabaleswar Panda wrote:
> > Hi Mark,
> > 
> > Thanks for providing us the details. There appears to be some
> > `time out' with the new mpirun_rsh. We are taking a look at it
> > and will be able to send you some solution soon.
> > 
> > Thanks,
> > 
> > DK
> > 
> > On Tue, 28 Aug 2007, Mark Potts wrote:
> > 
> >> Hi,
> >>     I tried VIADEV_USE_SHMEM_COLL=0 and separately tried
> >>     VIADEV_USE_BLOCKING=1, with no change in results.  During task
> >>     startup I get either "Unable to find child nnnn!", "Child died.
> >>     Timeout while waiting", and/or simply "done."
> >>
> >>     I tried repeatedly but was never able to consistently run more
> >>     than 10 ranks (-np 10) on a single node.  I, of course, am
> >>     able to run many more ranks, when I spread the targets across
> >>     more nodes.
> >>
> >>    My experiment is to start a very simple code with multiple
> >>    processes on a single node.  Specific details of my setup on two
> >>    machines.  The results were the same:
> >>
> >>     Machine Cpus per  Cores per   Avail   Cpu      MVAPICH     MVAPICH
> >>              Node      Cpu        Nodes   Type     version     Device
> >>      A        1         2          3     X86-64  -0.9.9-1168   ch_gen2
> >>      B        2         4         16     X86_64  -0.9.9-1326   ch_gen2
> >>
> >>     The MVAPICH code, which was obtained from ofed 1.2 installation,
> >>     has two patches as follows:
> >>      (1) for mpirun_rsh.c from Sayatan Sur of 10 Jul for MVAPICH errant
> >>          process/job cleanup.
> >>      (2) for comm_free.c from Amith Rajith Mamidla of 11 Jul for MVAPICH
> >>          segmentation fault during MPI_Finalize() in large jobs.
> >>
> >>     Is it possible that the mpirun_rsh.c patch is prematurely killing
> >>     tasks when it determines that the processes on the oversubscribed
> >>     node are not responding fast enough?  Or is there another
> >>     clean explanation?  As I understand the note from DK this morning,
> >>     oversubscription should work...
> >>           regards,
> >>
> >> amith rajith mamidala wrote:
> >>> Hi Mark,
> >>>
> >>> Can you check if you get this error by setting the environment variable:
> >>> VIADEV_USE_SHMEM_COLL to 0 e.g. mpirun_rsh -np N VIADEV_USE_SHMEM_COLL=0
> >>> ./a.out
> >>>
> >>> -thanks,
> >>> Amith
> >>>
> >>> On Tue, 28 Aug 2007, Mark Potts wrote:
> >>>
> >>>> Hi,
> >>>>     Is there an effective or hard limit on the number of MVAPICH
> >>>>     processes that can be run on a single node?
> >>>>
> >>>>     Given N cpus, each having M cores, on a single node, I've been told
> >>>>     that one can not run more than N*M MVAPICH processes on a single
> >>>>     node.  In fact, I observe that if I try to even approach this number
> >>>>     with "-np 16" (for a node with N=8 and M=4), I observe a "unable to
> >>>>     find child nnnn!" or "Child died" message.  Is this a configuration
> >>>>     problem with this system or somehow an expected behavior?
> >>>>
> >>>>     More pointedly, should oversubscription of cores, np > N*M, on a
> >>>>     single node work in MVAPICH?  How about in MVAPICH2?
> >>>>
> >>>>             regards,
> >>>> --
> >>>> ***********************************
> >>>>  >> Mark J. Potts, PhD
> >>>>  >>
> >>>>  >> HPC Applications Inc.
> >>>>  >> phone: 410-992-8360 Bus
> >>>>  >>        410-313-9318 Home
> >>>>  >>        443-418-4375 Cell
> >>>>  >> email: potts at hpcapplications.com
> >>>>  >>        potts at excray.com
> >>>> ***********************************
> >>>> _______________________________________________
> >>>> mvapich-discuss mailing list
> >>>> mvapich-discuss at cse.ohio-state.edu
> >>>> http://mail.cse.ohio-state.edu/mailman/listinfo/mvapich-discuss
> >>>>
> >> --
> >> ***********************************
> >>  >> Mark J. Potts, PhD
> >>  >>
> >>  >> HPC Applications Inc.
> >>  >> phone: 410-992-8360 Bus
> >>  >>        410-313-9318 Home
> >>  >>        443-418-4375 Cell
> >>  >> email: potts at hpcapplications.com
> >>  >>        potts at excray.com
> >> ***********************************
> >> _______________________________________________
> >> mvapich-discuss mailing list
> >> mvapich-discuss at cse.ohio-state.edu
> >> http://mail.cse.ohio-state.edu/mailman/listinfo/mvapich-discuss
> >>
> 
> -- 
> ***********************************
>  >> Mark J. Potts, PhD
>  >>
>  >> HPC Applications Inc.
>  >> phone: 410-992-8360 Bus
>  >>        410-313-9318 Home
>  >>        443-418-4375 Cell
>  >> email: potts at hpcapplications.com
>  >>        potts at excray.com
> ***********************************
> 



More information about the mvapich-discuss mailing list