[mvapich-discuss] Segmentation fault when compiling with Intel 15, mvapich2 2.1, and openmp

Craig Tierney - NOAA Affiliate craig.tierney at noaa.gov
Wed Oct 21 19:05:53 EDT 2015


Jonathan,

I will give this a try tomorrow.

Thanks,
Craig

On Wed, Oct 21, 2015 at 4:26 PM, Jonathan Perkins <
perkinjo at cse.ohio-state.edu> wrote:

> Hi Craig.  We haven't seen this issue with the Intel 2016 compilers.  Can
> you try this out?
>
>
> On Wed, Oct 21, 2015 at 1:07 PM Craig Tierney - NOAA Affiliate <
> craig.tierney at noaa.gov> wrote:
>
>> Jonathan,
>>
>> Are there any updates to this issue?  Intel's last response to this was
>> on 5/13 and they say it is an mvapich2 issue, not Intel.
>>
>> Craig
>>
>> On Thu, Sep 10, 2015 at 3:23 PM, Jonathan Perkins <
>> perkinjo at cse.ohio-state.edu> wrote:
>>
>>> Hi Craig.
>>>
>>> There has been extensive discussion about this issue between TACC and
>>> Intel. To the best of our
>>> knowledge, the issue is been localized to the Intel compiler.  This
>>> seems to affect versions 15.0.2 and up.
>>>
>>> On Thu, Sep 10, 2015 at 4:07 PM, Craig Tierney - NOAA Affiliate <
>>> craig.tierney at noaa.gov> wrote:
>>>
>>>> --===============3875389897677332204==
>>>> Content-Type: multipart/alternative;
>>>> boundary="047d7bdca5c272ca21051f6a262e"
>>>>
>>>> --047d7bdca5c272ca21051f6a262e
>>>> Content-Type: text/plain; charset="UTF-8"
>>>>
>>>>
>>>> Hello,
>>>>
>>>> I am having a problem building and running codes with Intel 15.X and
>>>> Mvapich2 2.1 that use openmp.  The following code shows the problem:
>>>>
>>>> program test
>>>>     use mpi
>>>>     integer :: ierr, iprov
>>>>     real(8) :: a
>>>>     call mpi_init_thread(MPI_THREAD_FUNNELED,iprov,ierr)
>>>>     call random_number(a)
>>>>     write(*,*)"hello"
>>>>     call mpi_finalize(ierr)
>>>> end program test
>>>>
>>>> If this is compiled as:
>>>>
>>>> # mpif90 -openmp -O3 test.f90 -o test
>>>>
>>>> And run, the following error message is generated.
>>>>
>>>> # mpiexec.hydra -np 2 ./testmpifort
>>>>
>>>>
>>>> ===================================================================================
>>>> =   BAD TERMINATION OF ONE OF YOUR APPLICATION PROCESSES
>>>> =   PID 17905 RUNNING AT fe8
>>>> =   EXIT CODE: 11
>>>> =   CLEANING UP REMAINING PROCESSES
>>>> =   YOU CAN IGNORE THE BELOW CLEANUP MESSAGES
>>>>
>>>> ===================================================================================
>>>> YOUR APPLICATION TERMINATED WITH THE EXIT STRING: Segmentation fault
>>>> (signal 11)
>>>> This typically refers to a problem with your application.
>>>> Please see the FAQ page for debugging suggestions
>>>>
>>>> When you get the stack trace, you see:
>>>>
>>>> (gdb) bt
>>>> #0  0x000000000040b0c4 in init_resource ()
>>>> #1  0x000000000040b02a in reentrancy_init ()
>>>> #2  0x000000000040af48 in for__reentrancy_init ()
>>>> #3  0x00002aaaab329115 in for_rtl_init_ () from
>>>> /home/admin/software/apps/mvapich2/2.1-intel/lib/libmpi.so.12
>>>> #4  0x0000000000403249 in main ()
>>>>
>>>> Intel believes this is a mvapich2 issue, not a compiler issue.
>>>>
>>>>
>>>> https://software.intel.com/en-us/forums/intel-fortran-compiler-for-linux-and-mac-os-x/topic/540673
>>>>
>>>> Thanks,
>>>> Craig
>>>>
>>>> --047d7bdca5c272ca21051f6a262e
>>>> Content-Type: text/html; charset="UTF-8"
>>>> Content-Transfer-Encoding: quoted-printable
>>>>
>>>> <div dir=3D"ltr">Hello,<div><br></div><div>I am having a problem
>>>> building a=
>>>> nd running codes with Intel 15.X and Mvapich2 2.1 that use
>>>> openmp.=C2=A0 Th=
>>>> e following code shows the
>>>> problem:</div><div><br></div><div><div>program t=
>>>> est</div><div>=C2=A0 =C2=A0 use mpi</div><div>=C2=A0 =C2=A0 integer ::
>>>> ierr=
>>>> , iprov</div><div>=C2=A0 =C2=A0 real(8) :: a</div><div>=C2=A0 =C2=A0
>>>> call m=
>>>> pi_init_thread(MPI_THREAD_FUNNELED,iprov,ierr)</div><div>=C2=A0 =C2=A0
>>>> call=
>>>>  random_number(a)</div><div>=C2=A0 =C2=A0
>>>> write(*,*)"hello"</div>=
>>>> <div>=C2=A0 =C2=A0 call mpi_finalize(ierr)</div><div>end program
>>>> test</div>=
>>>> </div><div><br></div><div>If this is compiled
>>>> as:</div><div><br></div><div>=
>>>> # mpif90 -openmp -O3 test.f90 -o test</div><div><br></div><div>And run,
>>>> the=
>>>>  following error message is generated.</div><div><br></div><div>#
>>>> mpiexec.h=
>>>> ydra -np 2
>>>> ./testmpifort</div><div><br></div><div>=3D=3D=3D=3D=3D=3D=3D=3D=
>>>>
>>>> =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=
>>>>
>>>> =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=
>>>>
>>>> =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=
>>>> </div><div>=3D =C2=A0 BAD TERMINATION OF ONE OF YOUR APPLICATION
>>>> PROCESSES<=
>>>> /div><div>=3D =C2=A0 PID 17905 RUNNING AT fe8</div><div>=3D =C2=A0 EXIT
>>>> COD=
>>>> E: 11</div><div>=3D =C2=A0 CLEANING UP REMAINING
>>>> PROCESSES</div><div>=3D =
>>>> =C2=A0 YOU CAN IGNORE THE BELOW CLEANUP
>>>> MESSAGES</div><div>=3D=3D=3D=3D=3D=
>>>>
>>>> =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=
>>>>
>>>> =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=
>>>>
>>>> =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=
>>>> =3D=3D=3D</div><div>YOUR APPLICATION TERMINATED WITH THE EXIT STRING:
>>>> Segme=
>>>> ntation fault (signal 11)</div><div>This typically refers to a problem
>>>> with=
>>>>  your application.</div><div>Please see the FAQ page for debugging
>>>> suggesti=
>>>> ons</div><div><br></div><div>When you get the stack trace, you
>>>> see:</div><d=
>>>> iv><br></div><div><div>(gdb) bt</div><div>#0 =C2=A00x000000000040b0c4
>>>> in in=
>>>> it_resource ()</div><div>#1 =C2=A00x000000000040b02a in reentrancy_init
>>>> ()<=
>>>> /div><div>#2 =C2=A00x000000000040af48 in for__reentrancy_init
>>>> ()</div><div>=
>>>> #3 =C2=A00x00002aaaab329115 in for_rtl_init_ () from
>>>> /home/admin/software/a=
>>>> pps/mvapich2/2.1-intel/lib/libmpi.so.12</div><div>#4
>>>> =C2=A00x00000000004032=
>>>> 49 in main ()</div></div><div><br></div><div>Intel believes this is a
>>>> mvapi=
>>>> ch2 issue, not a compiler issue.</div><div><br></div><div><a
>>>> href=3D"https:=
>>>> //
>>>> software.intel.com/en-us/forums/intel-fortran-compiler-for-linux-and-mac-=
>>>> os-x/topic/540673
>>>> <http://software.intel.com/en-us/forums/intel-fortran-compiler-for-linux-and-mac-=os-x/topic/540673>
>>>> ">https://software.intel.com/en-us/forums/intel-fortran-co=
>>>> mpiler-for-linux-and-mac-os-x/topic/540673
>>>> <https://software.intel.com/en-us/forums/intel-fortran-co=mpiler-for-linux-and-mac-os-x/topic/540673>
>>>> </a><br></div><div><br></div><div=
>>>> >Thanks,</div><div>Craig</div></div>
>>>>
>>>> --047d7bdca5c272ca21051f6a262e--
>>>>
>>>> --===============3875389897677332204==
>>>> Content-Type: text/plain; charset="us-ascii"
>>>> MIME-Version: 1.0
>>>> Content-Transfer-Encoding: 7bit
>>>> Content-Disposition: inline
>>>>
>>>> _______________________________________________
>>>> mvapich-discuss mailing list
>>>> mvapich-discuss at cse.ohio-state.edu
>>>> http://mailman.cse.ohio-state.edu/mailman/listinfo/mvapich-discuss
>>>>
>>>> --===============3875389897677332204==--
>>>>
>>>
>>>
>>>
>>> --
>>> Jonathan Perkins
>>> http://www.cse.ohio-state.edu/~perkinjo
>>>
>>
>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.cse.ohio-state.edu/pipermail/mvapich-discuss/attachments/20151021/6a352eff/attachment-0001.html>


More information about the mvapich-discuss mailing list