[mvapich-discuss] mvapich218a1p1 Error while reading PMI socket

bright.yang at vaisala.com bright.yang at vaisala.com
Wed Nov 23 11:59:35 EST 2011


Here is the link to an old discussion related 
http://mail.cse.ohio-state.edu/pipermail/mvapich-discuss/2011-August/003507.html
Ibstat and mpiname are not found in my system. This is a new system so I suspect there are software needed to install. Here is the output of ldd wrf.exe
# ldd wrf.exe
        linux-vdso.so.1 =>  (0x00007fff1bdce000)
        libpsm_infinipath.so.1 => /usr/lib64/libpsm_infinipath.so.1 (0x00002af4a184c000)
        libjasper.so.1 => /usr/lib64/libjasper.so.1 (0x000000399a600000)
        libmpichf90.so.3 => /share/apps/mvapich218a1p1/lib/libmpichf90.so.3 (0x00002af4a1a9a000)
        libmpich.so.3 => /share/apps/mvapich218a1p1/lib/libmpich.so.3 (0x00002af4a1c9c000)
        libopa.so.1 => /share/apps/mvapich218a1p1/lib/libopa.so.1 (0x00002af4a2027000)
        libmpl.so.1 => /share/apps/mvapich218a1p1/lib/libmpl.so.1 (0x00002af4a2229000)
        libpthread.so.0 => /lib64/libpthread.so.0 (0x00000032cae00000)
        librt.so.1 => /lib64/librt.so.1 (0x00000032cb600000)
        libm.so.6 => /lib64/libm.so.6 (0x00000032ca600000)
        libc.so.6 => /lib64/libc.so.6 (0x00000032ca200000)
        libinfinipath.so.4 => /usr/lib64/libinfinipath.so.4 (0x00002af4a242e000)
        libdl.so.2 => /lib64/libdl.so.2 (0x00000032caa00000)
        libjpeg.so.62 => /usr/lib64/libjpeg.so.62 (0x00000032d5000000)
        /lib64/ld-linux-x86-64.so.2 (0x00000032c9e00000)

Bright Yang

-----Original Message-----
From: Jonathan Perkins [mailto:perkinjo at cse.ohio-state.edu] 
Sent: Wednesday, November 23, 2011 9:43 AM
To: Yang Bright BRYA
Cc: mvapich-discuss at cse.ohio-state.edu
Subject: Re: [mvapich-discuss] mvapich218a1p1 Error while reading PMI socket

Hello, I've done a quick search but I'm not sure which issue this may
correspond to.  Can you send me the output of mpiname -a as well as
ldd wrf.exe and ibstat?  Thanks in advance.

On Wed, Nov 23, 2011 at 11:23 AM,  <bright.yang at vaisala.com> wrote:
> Hi,
>
>
>
>   I got this error message while running mpirun_rsh. I saw some old message
> on the same error but I can't find solution for it. -
>
> $ mpirun_rsh -np 24 -hostfile hosts57 wrf.exe
>
>   starting wrf task             1   of starting wrf task            24
>
>              3 starting wrf task  starting wrf task  starting wrf task
> starting wrf task  starting wrf task   of   starting wrf task starting wrf
> task      starting wrf task           21starting wrf task           23
> starting wrf task     of  of           22starting wrf task
> 18             11              24            9            7            10
> starting wrf task            6           24  starting wrf task
>
> starting wrf task starting wrf task            24 of           17
> 19 of
>
> starting wrf task
>
>            13           24
>
>  starting wrf task  of           24starting wrf task
>
> starting wrf task  of      of            5           4             24 of
> starting wrf task    of            24
>
>   of              24
>
>  of           24
>
>            24          15starting wrf task
>
>  starting wrf task           20 of                16 of
>   12          24          14              24 of            24
> of            2
>
>  of
>
>
>
>              0          24 of           24             8 of            24
>
>            24
>
>           24
>
>
>
>  of            24
>
>
>
>  of           24 of  of
>
>
>
>            24          24          24
>
>
>
>
>
> [compute-0-5.local:mpispawn_0][readline] Unexpected End-Of-File on file
> descriptor 10. MPI process died?
>
> [compute-0-5.local:mpispawn_0][mtpmi_processops] Error while reading PMI
> socket. MPI process died?
>
> [compute-0-7.local:mpispawn_1][readline] Unexpected End-Of-File on file
> descriptor 7. MPI process died?
>
> [compute-0-7.local:mpispawn_1][mtpmi_processops] Error while reading PMI
> socket. MPI process died?
>
> [compute-0-7.local:mpispawn_1][child_handler] MPI process (rank: 15, pid:
> 5330) exited with status 1
>
> [compute-0-7.local:mpispawn_1][child_handler] MPI process (rank: 22, pid:
> 5337) exited with status 1
>
> [compute-0-7.local:mpispawn_1][child_handler] MPI process (rank: 23, pid:
> 5338) exited with status 1
>
> [compute-0-7.local:mpispawn_1][child_handler] MPI process (rank: 19, pid:
> 5334) exited with status 1
>
> [compute-0-7.local:mpispawn_1][child_handler] MPI process (rank: 13, pid:
> 5328) exited with status 1
>
> [compute-0-7.local:mpispawn_1][child_handler] MPI process (rank: 12, pid:
> 5327) exited with status 1
>
> [compute-0-7.local:mpispawn_1][child_handler] MPI process (rank: 20, pid:
> 5335) exited with status 1
>
> [compute-0-7.local:mpispawn_1][child_handler] MPI process (rank: 18, pid:
> 5333) exited with status 1
>
> [compute-0-7.local:mpispawn_1][child_handler] MPI process (rank: 17, pid:
> 5332) exited with status 1
>
> [compute-0-7.local:mpispawn_1][child_handler] MPI process (rank: 16, pid:
> 5331) exited with status 1
>
> [compute-0-7.local:mpispawn_1][child_handler] MPI process (rank: 21, pid:
> 5336) exited with status 1
>
> [compute-0-5.local:mpispawn_0][child_handler] MPI process (rank: 5, pid:
> 11421) exited with status 1
>
> [compute-0-5.local:mpispawn_0][child_handler] MPI process (rank: 7, pid:
> 11423) exited with status 1
>
> [compute-0-7.local:mpispawn_1][child_handler] MPI process (rank: 14, pid:
> 5329) exited with status 1
>
> [compute-0-5.local:mpispawn_0][child_handler] MPI process (rank: 1, pid:
> 11417) exited with status 1
>
> [compute-0-5.local:mpispawn_0][child_handler] MPI process (rank: 4, pid:
> 11420) exited with status 1
>
> [compute-0-5.local:mpispawn_0][child_handler] MPI process (rank: 0, pid:
> 11416) exited with status 1
>
> [compute-0-5.local:mpispawn_0][child_handler] MPI process (rank: 3, pid:
> 11419) exited with status 1
>
> [compute-0-5.local:mpispawn_0][child_handler] MPI process (rank: 6, pid:
> 11422) exited with status 1
>
> [compute-0-5.local:mpispawn_0][child_handler] MPI process (rank: 11, pid:
> 11427) exited with status 1
>
> [compute-0-5.local:mpispawn_0][child_handler] MPI process (rank: 10, pid:
> 11426) exited with status 1
>
> [compute-0-5.local:mpispawn_0][child_handler] MPI process (rank: 9, pid:
> 11425) exited with status 1
>
> [compute-0-5.local:mpispawn_0][child_handler] MPI process (rank: 2, pid:
> 11418) exited with status 1
>
> [compute-0-5.local:mpispawn_0][child_handler] MPI process (rank: 8, pid:
> 11424) exited with status 1
>
>
>
> Thanks.
>
>
>
> Bright Yang
>
> Vaisala, Boulder
>
> _______________________________________________
> mvapich-discuss mailing list
> mvapich-discuss at cse.ohio-state.edu
> http://mail.cse.ohio-state.edu/mailman/listinfo/mvapich-discuss
>
>



-- 
Jonathan Perkins
http://www.cse.ohio-state.edu/~perkinjo



More information about the mvapich-discuss mailing list