[mvapich-discuss] mpiexec/mvapich places processes on same cpu

LEI CHAI chai.15 at osu.edu
Sun Jan 20 20:20:38 EST 2008


Hi Joseph and Pasha,

Thank you for the suggestions. We already plan to add this feature to MVAPICH. It will be available in upcoming releases.

Lei


----- Original Message -----
From: "Pavel Shamis (Pasha)" <pasha at dev.mellanox.co.il>
Date: Sunday, January 20, 2008 4:45 am
Subject: Re: [mvapich-discuss] mpiexec/mvapich places processes on same cpu

> 
> > While we are at it - is there an actual way to assign with 
> either version to specific cores? We have a few large multinode 
> obs that can only use 4 out of 8 cores per node due to bus and 
> memory limitations. Is there a way to distribute the processes to 
> cores 1,2 and 5,6 ? ie to skip 0, and split the other 4 to 
> different chips? 
> >   
> Sounds good. As we have a option to specify HCA/port per rank it 
> may be 
> nice to have option to specify core. Should not be very 
> complicated to 
> implement. (I'm talking about mvapich1)
> Mvapich team, what do you think ?
> 
> Pasha
> > j
> >
> > ----- Original Message -----
> > From: LEI CHAI <chai.15 at osu.edu>
> > Date: Friday, January 18, 2008 7:00 pm
> > Subject: Re: [mvapich-discuss] mpiexec/mvapich places processes 
> on same cpu
> >
> >   
> >> Hi Joseph,
> >>
> >> Could you try disable the cpu affinity feature in 
> mvapich/mvapich2, e.g.
> >>
> >> mvapich2:
> >> $ mpiexec -n 4 -env MV2_ENABLE_AFFINITY 0 ./a.out
> >>
> >> or mvapich:
> >> $ mpirun_rsh -np 4 VIADEV_ENABLE_AFFINITY=0 ./a.out
> >>
> >> Thanks,
> >> Lei
> >>
> >>
> >> ----- Original Message -----
> >> From: Joseph Hargitai <joseph.hargitai at nyu.edu>
> >> Date: Friday, January 18, 2008 6:05 pm
> >> Subject: [mvapich-discuss] mpiexec/mvapich places processes on 
> same cpu
> >>
> >>     
> >>> hi all:
> >>>
> >>> While submitting two identical mpi jobs (-np 4) with 
> differetnt 
> >>> datasets for a dual socket quadocore node using two distinct 
> >>> pbs/mpiexec submission both jobs end up on the first 
> processor, 
> >>> such they use 4 cores of the first cpu, none of the second. 
> This 
> >>> results in 8 processes on cpu 1, with a load of about 8-9, 
> both 
> >>> jobs producing output okay, but obviously the choice would be 
> to 
> >>> have them on distinct cpus. 
> >>>
> >>> When one of these mpi jobs meet 4 others regular serial jobs 
> >>> submitted without mpiexec, all 8 cores are populated. 
> >>>
> >>> I did read on the group list about the first mpiexec session 
> being 
> >>> the master, but not reserving the first 4 cores such allowing 
> the 
> >>> possibility for the next mpi job to end up on the same cpu. 
> >>>
> >>> best,   
> >>> joseph
> >>> _______________________________________________
> >>> mvapich-discuss mailing list
> >>> mvapich-discuss at cse.ohio-state.edu
> >>> http://mail.cse.ohio-state.edu/mailman/listinfo/mvapich-discuss
> >>>
> >>>       
> > _______________________________________________
> > mvapich-discuss mailing list
> > mvapich-discuss at cse.ohio-state.edu
> > http://mail.cse.ohio-state.edu/mailman/listinfo/mvapich-discuss
> >
> >   
> 
> 
> -- 
> Pavel Shamis (Pasha)
> Mellanox Technologies
> 
> 



More information about the mvapich-discuss mailing list