[Mvapich-discuss] [mvapich-discuss] MVAPICH2 2.3.5 and HDF5 1.10.7 parallel tests
Subramoni, Hari
subramoni.1 at osu.edu
Thu Mar 18 11:49:56 EDT 2021
Hi, All.
We had some offline discussion on this. We were able to resolve the issue. The solution will be available in the next release of MVAPICH2.
Best,
Hari.
-----Original Message-----
From: Subramoni, Hari <subramoni.1 at osu.edu>
Sent: Tuesday, December 15, 2020 1:42 PM
To: Mark Dixon <mark.c.dixon at durham.ac.uk>; mvapich-discuss at cse.ohio-state.edu <mvapich-discuss at mailman.cse.ohio-state.edu>
Cc: Subramoni, Hari <subramoni.1 at osu.edu>
Subject: RE: [mvapich-discuss] MVAPICH2 2.3.5 and HDF5 1.10.7 parallel tests
Hi, Mark.
Sorry to hear that you're facing issues.
Can you please let us know if the is issue particular to POWER9 + Lustre + UFS combination or does it happen on x86 systems as well?
We will try out the steps you've mentioned locally and see if we are able to reproduce it.
Thx,,
Hari.
-----Original Message-----
From: mvapich-discuss-bounces at cse.ohio-state.edu <mvapich-discuss-bounces at mailman.cse.ohio-state.edu> On Behalf Of Mark Dixon
Sent: Tuesday, December 15, 2020 12:11 PM
To: mvapich-discuss at cse.ohio-state.edu <mvapich-discuss at mailman.cse.ohio-state.edu>
Subject: [mvapich-discuss] MVAPICH2 2.3.5 and HDF5 1.10.7 parallel tests
Hi there,
I'm having trouble getting HDF5's parallel tests to pass when built on top of MVAPICH2. I was wondering if anyone else is seeing this, please?
For reference (not sure it's relevant), I had similar trouble with the version of ROMIO bundled inside OpenMPI (https://urldefense.com/v3/__https://github.com/open-mpi/ompi/issues/6871__;!!KGKeukY!nWOQssfAxiSnSZJJ7bf62O6vhk438EzOmw7yXziKQ3niAaJZUYlmw-6k6nKp0Si8pwjiqKIs-3I4CJc$ )
Thanks,
Mark
#!/bin/bash
# We have run this on an IBM POWER9 rhel7.6-alt system with MOFED 4.7. The # hdf5 test "testphdf5" does not complete until it is terminated by (in # this case) a 1 hour alarm timeout.
#
# The last lines that the test printed was 6 copies of this:
#
# Testing -- multi-chunk collective chunk io (cchunk3) # # This has been run from a location on an xfs filesystem, and # on a lustre filesystem, with the same result.
set -x
set -e
# (needed on our system to ensure we are using the OS-provided # version of GCC, etc.) module purge || true
test -d build || mkdir build
test -d src || mkdir src
prefix=`pwd`/build
export PATH=${prefix}/bin:$PATH
cd src
# mvapich2
wget https://mvapich.cse.ohio-state.edu/download/mvapich/mv2/mvapich2-2.3.5.tar.gz
tar xf mvapich2-2.3.5.tar.gz
(
cd mvapich2-2.3.5
./configure --prefix=$prefix \
--enable-romio \
--with-file-system=lustre+ufs
make -j12
make install
)
# hdf5
wget https://urldefense.com/v3/__https://support.hdfgroup.org/ftp/HDF5/releases/hdf5-1.10/hdf5-1.10.7/src/hdf5-1.10.7.tar.gz__;!!KGKeukY!nWOQssfAxiSnSZJJ7bf62O6vhk438EzOmw7yXziKQ3niAaJZUYlmw-6k6nKp0Si8pwjiqKIsvYU7Edc$
tar xf hdf5-1.10.7.tar.gz
(
cd hdf5-1.10.7
export CC=mpicc
export CXX=mpicxx
export FC=mpif90
export F77=mpif77
export F90=mpif90
export HDF5_ALARM_SECONDS=3600
./configure --prefix=$prefix --enable-parallel
make -j12
make check
make install
)
_______________________________________________
mvapich-discuss mailing list
mvapich-discuss at cse.ohio-state.edu
http://mailman.cse.ohio-state.edu/mailman/listinfo/mvapich-discuss
More information about the Mvapich-discuss
mailing list