[mvapich-discuss] No route to host in MVAPICH-Virt

Maksym Planeta mplaneta at os.inf.tu-dresden.de
Wed Jul 5 04:54:40 EDT 2017



On 07/04/2017 06:05 PM, Jie Zhang wrote:
> Hi,  Maksym,
> 
> This should be a network setup issue.
> Can you please try a few things:
> 
> 1. Do you have correct IP/hostname mapping for both hosts in the file,
> such as /etc/hosts 

Yes, both /etc/hosts and /etc/hostname are consistent on both VMs.

> 2. What is the output if you run /opt/mvapich2/virt/2.2/bin/mpirun_rsh 
> -np 2 host11 host12 MV2_VIRT_USE_IVSHMEM=1 hostname
> 


[user at host12 mpi_init]$ /opt/mvapich2/virt/2.2/bin/mpirun_rsh  -np 2 host11 host12 MV2_VIRT_USE_IVSHMEM=1 hostname
host12
connect [mt_checkin]: No route to host
[host12:mpirun_rsh][child_handler] Error in init phase, aborting! (1/2 mpispawn connections)



If it helps, I can give you output of ip addr:

 ip addr
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN qlen 1
    link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
    inet 127.0.0.1/8 scope host lo
       valid_lft forever preferred_lft forever
    inet6 ::1/128 scope host        valid_lft forever preferred_lft forever
2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state UP qlen 1000
    link/ether 52:54:00:6b:c7:ee brd ff:ff:ff:ff:ff:ff
    inet 141.76.84.2/26 brd 141.76.84.63 scope global eth0
       valid_lft forever preferred_lft forever
    inet6 2002:8d4c:3001:48:ef38:8a6:5326:dd4d/64 scope global noprefixroute dynamic        valid_lft 86296sec preferred_lft 14296sec
    inet6 fe80::c092:6007:b126:618b/64 scope link        valid_lft forever preferred_lft forever
3: eth1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state UP qlen 1000
    link/ether 52:54:00:69:81:11 brd ff:ff:ff:ff:ff:ff
    inet 192.168.254.2/24 brd 192.168.254.255 scope global dynamic eth1
       valid_lft 3016sec preferred_lft 3016sec
    inet6 fe80::984e:1704:c529:5607/64 scope link        valid_lft forever preferred_lft forever

The eth1 interface is an interface which allows me to ssh from the host into the VM. I created it following this advice:

https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_Linux/6/html/Virtualization_Host_Configuration_and_Guest_Installation_Guide/App_Macvtap.html

It looks similarly for the second host.

> 
> On Tue, Jul 4, 2017 at 11:01 AM Maksym Planeta
> <mplaneta at os.inf.tu-dresden.de <mailto:mplaneta at os.inf.tu-dresden.de>>
> wrote:
> 
>     Hello,
> 
>     I have setup several virtual machines and I'm trying to run simple
>     MPI programs in there.
> 
>     My current test is following:
> 
>     /opt/mvapich2/virt/2.2/bin/mpirun_rsh  -export-all -n 2 host11
>     host12 MV2_VIRT_USE_IVSHMEM=1 bash -c 'netstat -l | grep $(printenv
>     MPISPAWN_CHECKIN_PORT)'
>     tcp        0      0 0.0.0.0:49379 <http://0.0.0.0:49379>         
>      0.0.0.0:*               LISTEN
>     connect [mt_checkin]: No route to host
>     [host11:mpirun_rsh][child_handler] Error in init phase, aborting!
>     (1/2 mpispawn connections)
> 
>     So I'm trying to look if the checkin port is open. But the remote
>     host connection fails with a message "no route to host".
> 
>     Showing the command which is about to be executed gives me this:
> 
>     $ /opt/mvapich2/virt/2.2/bin/mpirun_rsh -show  -export-all -n 2
>     host11 host12 MV2_VIRT_USE_IVSHMEM=1 bash -c 'netstat -l | grep
>     $(printenv MPISPAWN_CHECKIN_PORT)'
>     /usr/bin/ssh -q host12 cd /home/user; /usr/bin/env 
>     MPISPAWN_MPIRUN_MPD=0 USE_LINEAR_SSH=1 MPISPAWN_MPIRUN_HOST=host11
>     MPISPAWN_MPIRUN_HOSTIP=141.76.84.15 MPIRUN_RSH_LAUNCH=1
>     MPISPAWN_CHECKIN_PORT=52056 MPISPAWN_MPIRUN_PORT=52056
>     MPISPAWN_NNODES=2 MPISPAWN_GLOBAL_NPROCS=2 MPISPAWN_MPIRUN_ID=5190
>     MPISPAWN_ARGC=3 MPDMAN_KVS_TEMPLATE=kvs_506_host11_5190
>     MPISPAWN_LOCAL_NPROCS=1 MPISPAWN_ARGV_0='bash' MPISPAWN_ARGV_1='-c'
>     MPISPAWN_ARGV_2='netstat -l | grep $(printenv
>     MPISPAWN_CHECKIN_PORT)' MPISPAWN_ARGC=3
>     MPISPAWN_GENERIC_ENV_COUNT=1 
>     MPISPAWN_GENERIC_NAME_0=MV2_VIRT_USE_IVSHMEM
>     MPISPAWN_GENERIC_VALUE_0=1 MPISPAWN_ID=1
>     MPISPAWN_WORKING_DIR=/home/user MPISPAWN_MPIRUN_RANK_0=1 
>     /opt/mvapich2/virt/2.2/bin/mpispawn 0
> 
>     /usr/bin/bash -c cd /home/user; /usr/bin/env  MPISPAWN_MPIRUN_MPD=0
>     USE_LINEAR_SSH=1 MPISPAWN_MPIRUN_HOST=host11
>     MPISPAWN_MPIRUN_HOSTIP=141.76.84.15 MPIRUN_RSH_LAUNCH=1
>     MPISPAWN_CHECKIN_PORT=52056 MPISPAWN_MPIRUN_PORT=52056
>     MPISPAWN_NNODES=2 MPISPAWN_GLOBAL_NPROCS=2 MPISPAWN_MPIRUN_ID=5190
>     MPISPAWN_ARGC=3 MPDMAN_KVS_TEMPLATE=kvs_506_host11_5190
>     MPISPAWN_LOCAL_NPROCS=1 MPISPAWN_ARGV_0='bash' MPISPAWN_ARGV_1='-c'
>     MPISPAWN_ARGV_2='netstat -l | grep $(printenv
>     MPISPAWN_CHECKIN_PORT)' MPISPAWN_ARGC=3
>     MPISPAWN_GENERIC_ENV_COUNT=1 
>     MPISPAWN_GENERIC_NAME_0=MV2_VIRT_USE_IVSHMEM
>     MPISPAWN_GENERIC_VALUE_0=1 MPISPAWN_ID=0
>     MPISPAWN_WORKING_DIR=/home/user MPISPAWN_MPIRUN_RANK_0=0 
>     /opt/mvapich2/virt/2.2/bin/mpispawn 0
> 
> 
>     I can reach both hosts from each other and they see the same IP
>     addresses, and I can ssh to each of the two hosts from each of the
>     two hosts.
> 
> 
>     Could you help me to identify the issue?
> 
>     The MVAPICH-Virt version is the last accessible on the mvapich
>     download page
> 
>     P. S.
> 
>     You say to allow reading for the user in section 5 of the Virt user
>     manual, but actually the library asks for write access as well, so I
>     set up rights to 766.
> 
>     --
>     Regards,
>     Maksym Planeta
> 
>     _______________________________________________
>     mvapich-discuss mailing list
>     mvapich-discuss at cse.ohio-state.edu
>     <mailto:mvapich-discuss at cse.ohio-state.edu>
>     http://mailman.cse.ohio-state.edu/mailman/listinfo/mvapich-discuss
> 
> -- 
> Thanks,
> -Jie
> 
> --------------------------------------------
> Department of Computer Science & Engineering
> Erwin E. Dreese Laboratories, 2015 Neil Avenue
> Ohio State University

-- 
Regards,
Maksym Planeta


-------------- next part --------------
A non-text attachment was scrubbed...
Name: smime.p7s
Type: application/pkcs7-signature
Size: 5174 bytes
Desc: S/MIME Cryptographic Signature
URL: <http://mailman.cse.ohio-state.edu/pipermail/mvapich-discuss/attachments/20170705/e00a15bc/attachment.p7s>


More information about the mvapich-discuss mailing list