[mvapich-discuss] mvapich2 1.9a fails to cleanup after failed jobs
Bhupender Thakur
bthakur at lsu.edu
Mon Nov 19 12:05:26 EST 2012
Hi,
We are working on implementing mvapich2 one out new cluster but have run into some issues
with mvapich2 unable to cleanup frequently when jobs fails
We are usign mellanox infiniband
$ ibv_devinfo
hca_id: mlx4_0
transport: InfiniBand (0)
fw_ver: 2.10.4492
node_guid: 0002:c903:00ff:25b0
sys_image_guid: 0002:c903:00ff:25b3
vendor_id: 0x02c9
vendor_part_id: 4099
hw_ver: 0x0
board_id: DEL0A30000019
phys_port_cnt: 1
port: 1
state: PORT_ACTIVE (4)
max_mtu: 2048 (4)
active_mtu: 2048 (4)
sm_lid: 1
port_lid: 300
port_lmc: 0x00
link_layer: IB
mvapich2 1.9a
$ mpiname -a
MVAPICH2 1.9a Sat Sep 8 15:01:35 EDT 2012 ch3:mrail
Compilation
CC: /usr/local/compilers/Intel/composer_xe_2013.0.079/bin/intel64/icc -O2 -fPIC -g -DNDEBUG -DNVALGRIND -O2
CXX: /usr/local/compilers/Intel/composer_xe_2013.0.079/bin/intel64/icpc -O2 -fPIC -g -DNDEBUG -DNVALGRIND -O2
F77: /usr/local/compilers/Intel/composer_xe_2013/bin/ifort -g -O2 -L/usr/lib64
FC: /usr/local/compilers/Intel/composer_xe_2013.0.079/bin/intel64/ifort -O2 -fPIC -g -O2
Configuration
--prefix=/usr/local/packages/mvapich2/1.9a/Intel-13.0.0 \
FC=/usr/local/compilers/Intel/composer_xe_2013.0.079/bin/intel64/ifort \
CC=/usr/local/compilers/Intel/composer_xe_2013.0.079/bin/intel64/icc \
CXX=/usr/local/compilers/Intel/composer_xe_2013.0.079/bin/intel64/icpc \
CFLAGS=-O2 -fPIC FCFLAGS=-O2 -fPIC CXXFLAGS=-O2 -fPIC \
LDFLAGS=-L/usr/local/compilers/Intel/composer_xe_2013/lib -L/usr/local/compilers/Intel/composer_xe_2013/lib/intel64 \
LIBS= CPPFLAGS= \
--enable-rdma-cm --enable-g=dbg --enable-romio --with-file-system=lustre+nfs \
--with-ib-include=/usr/include --with-ib-libpath=/usr/lib64 \
--enable-threads=runtime --enable-mpe --enable-smpcoll --enable-shared --enable-xrc --with-hwloc
$ pbs_mom --version
version: 3.0.6
pbs moms are threaded with the default 3 threads.
This does not happen with openmpi. A sample hello world is being run with different parameters
program dummy
use mpi
character*10 name
! Init MPI
call MPI_Init(mpierr)
! Get Rank Size
call MPI_COMM_Rank(MPI_COMM_WORLD, nrank, mpierr)
call MPI_COMM_Size(MPI_COMM_WORLD, nproc, mpierr)
! Get Date
if (nrank==0) then
write(*,*)'System date: Running mpirun_rsh'
call system('date')
end if
! Print rank
call MPI_Barrier(MPI_COMM_WORLD, mpierr)
!
call MPI_Get_processor_name(name, nlen, mpierr)
write(*,*)" I am ", nrank, " of " ,nproc, " on ", name
!
call MPI_Barrier(MPI_COMM_WORLD, mpierr)
! Finalize
call MPI_Finalize(mpierr)
end
===========
#
cat $PBS_NODEFILE > hostfile
cat $PBS_NODEFILE | uniq > hosts
mpi_width=`cat hostfile | wc -l`
mpi_nodes=`cat hosts | wc -l`
for mpi_pn in 8 16
do
let sum_mpi=$mpi_pn*$mpi_nodes
let OMP_NUM_THREADS=$mpi_width/$sum_mpi
for param in "MV2_USE_XRC=1" "MV2_USE_RoCE=1" "MV2_USE_RDMA_CM=1"
do
echo " $param"
echo " nodes:$mpi_nodes mpi-per-node:$mpi_pn omp:$OMP_NUM_THREADS"
time mpirun_rsh -np $sum_mpi -hostfile hosts $param ./dummy
done
done
===============
using parameter "MV2_USE_RoCE=1" and "MV2_USE_RDMA_CM=1" should fail as they
not been configured yet(openib.conf), nevertheless, the program does not exit cleanly.
We are seeing this with some other applications where the process seems to have crashed
and is not producting any useful output, but there are threads lingering on ling after the program has
crashed.
At this stage we are not sure if this is infiniband or torque or mvapich2 issue. Please let us know if
you have seen this behaviour and if there is a way to resolve this.
Best,
Bhupender.
Bhupender Thakur.
IT- Analyst,
High Performance Computing, LSU.
Ph (225)578-5934
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://mail.cse.ohio-state.edu/pipermail/mvapich-discuss/attachments/20121119/6244db91/attachment.html
More information about the mvapich-discuss
mailing list