[mvapich-discuss] No route to host in MVAPICH-Virt
Maksym Planeta
mplaneta at os.inf.tu-dresden.de
Wed Jul 5 04:54:40 EDT 2017
On 07/04/2017 06:05 PM, Jie Zhang wrote:
> Hi, Maksym,
>
> This should be a network setup issue.
> Can you please try a few things:
>
> 1. Do you have correct IP/hostname mapping for both hosts in the file,
> such as /etc/hosts
Yes, both /etc/hosts and /etc/hostname are consistent on both VMs.
> 2. What is the output if you run /opt/mvapich2/virt/2.2/bin/mpirun_rsh
> -np 2 host11 host12 MV2_VIRT_USE_IVSHMEM=1 hostname
>
[user at host12 mpi_init]$ /opt/mvapich2/virt/2.2/bin/mpirun_rsh -np 2 host11 host12 MV2_VIRT_USE_IVSHMEM=1 hostname
host12
connect [mt_checkin]: No route to host
[host12:mpirun_rsh][child_handler] Error in init phase, aborting! (1/2 mpispawn connections)
If it helps, I can give you output of ip addr:
ip addr
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN qlen 1
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
valid_lft forever preferred_lft forever
inet6 ::1/128 scope host valid_lft forever preferred_lft forever
2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state UP qlen 1000
link/ether 52:54:00:6b:c7:ee brd ff:ff:ff:ff:ff:ff
inet 141.76.84.2/26 brd 141.76.84.63 scope global eth0
valid_lft forever preferred_lft forever
inet6 2002:8d4c:3001:48:ef38:8a6:5326:dd4d/64 scope global noprefixroute dynamic valid_lft 86296sec preferred_lft 14296sec
inet6 fe80::c092:6007:b126:618b/64 scope link valid_lft forever preferred_lft forever
3: eth1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state UP qlen 1000
link/ether 52:54:00:69:81:11 brd ff:ff:ff:ff:ff:ff
inet 192.168.254.2/24 brd 192.168.254.255 scope global dynamic eth1
valid_lft 3016sec preferred_lft 3016sec
inet6 fe80::984e:1704:c529:5607/64 scope link valid_lft forever preferred_lft forever
The eth1 interface is an interface which allows me to ssh from the host into the VM. I created it following this advice:
https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_Linux/6/html/Virtualization_Host_Configuration_and_Guest_Installation_Guide/App_Macvtap.html
It looks similarly for the second host.
>
> On Tue, Jul 4, 2017 at 11:01 AM Maksym Planeta
> <mplaneta at os.inf.tu-dresden.de <mailto:mplaneta at os.inf.tu-dresden.de>>
> wrote:
>
> Hello,
>
> I have setup several virtual machines and I'm trying to run simple
> MPI programs in there.
>
> My current test is following:
>
> /opt/mvapich2/virt/2.2/bin/mpirun_rsh -export-all -n 2 host11
> host12 MV2_VIRT_USE_IVSHMEM=1 bash -c 'netstat -l | grep $(printenv
> MPISPAWN_CHECKIN_PORT)'
> tcp 0 0 0.0.0.0:49379 <http://0.0.0.0:49379>
> 0.0.0.0:* LISTEN
> connect [mt_checkin]: No route to host
> [host11:mpirun_rsh][child_handler] Error in init phase, aborting!
> (1/2 mpispawn connections)
>
> So I'm trying to look if the checkin port is open. But the remote
> host connection fails with a message "no route to host".
>
> Showing the command which is about to be executed gives me this:
>
> $ /opt/mvapich2/virt/2.2/bin/mpirun_rsh -show -export-all -n 2
> host11 host12 MV2_VIRT_USE_IVSHMEM=1 bash -c 'netstat -l | grep
> $(printenv MPISPAWN_CHECKIN_PORT)'
> /usr/bin/ssh -q host12 cd /home/user; /usr/bin/env
> MPISPAWN_MPIRUN_MPD=0 USE_LINEAR_SSH=1 MPISPAWN_MPIRUN_HOST=host11
> MPISPAWN_MPIRUN_HOSTIP=141.76.84.15 MPIRUN_RSH_LAUNCH=1
> MPISPAWN_CHECKIN_PORT=52056 MPISPAWN_MPIRUN_PORT=52056
> MPISPAWN_NNODES=2 MPISPAWN_GLOBAL_NPROCS=2 MPISPAWN_MPIRUN_ID=5190
> MPISPAWN_ARGC=3 MPDMAN_KVS_TEMPLATE=kvs_506_host11_5190
> MPISPAWN_LOCAL_NPROCS=1 MPISPAWN_ARGV_0='bash' MPISPAWN_ARGV_1='-c'
> MPISPAWN_ARGV_2='netstat -l | grep $(printenv
> MPISPAWN_CHECKIN_PORT)' MPISPAWN_ARGC=3
> MPISPAWN_GENERIC_ENV_COUNT=1
> MPISPAWN_GENERIC_NAME_0=MV2_VIRT_USE_IVSHMEM
> MPISPAWN_GENERIC_VALUE_0=1 MPISPAWN_ID=1
> MPISPAWN_WORKING_DIR=/home/user MPISPAWN_MPIRUN_RANK_0=1
> /opt/mvapich2/virt/2.2/bin/mpispawn 0
>
> /usr/bin/bash -c cd /home/user; /usr/bin/env MPISPAWN_MPIRUN_MPD=0
> USE_LINEAR_SSH=1 MPISPAWN_MPIRUN_HOST=host11
> MPISPAWN_MPIRUN_HOSTIP=141.76.84.15 MPIRUN_RSH_LAUNCH=1
> MPISPAWN_CHECKIN_PORT=52056 MPISPAWN_MPIRUN_PORT=52056
> MPISPAWN_NNODES=2 MPISPAWN_GLOBAL_NPROCS=2 MPISPAWN_MPIRUN_ID=5190
> MPISPAWN_ARGC=3 MPDMAN_KVS_TEMPLATE=kvs_506_host11_5190
> MPISPAWN_LOCAL_NPROCS=1 MPISPAWN_ARGV_0='bash' MPISPAWN_ARGV_1='-c'
> MPISPAWN_ARGV_2='netstat -l | grep $(printenv
> MPISPAWN_CHECKIN_PORT)' MPISPAWN_ARGC=3
> MPISPAWN_GENERIC_ENV_COUNT=1
> MPISPAWN_GENERIC_NAME_0=MV2_VIRT_USE_IVSHMEM
> MPISPAWN_GENERIC_VALUE_0=1 MPISPAWN_ID=0
> MPISPAWN_WORKING_DIR=/home/user MPISPAWN_MPIRUN_RANK_0=0
> /opt/mvapich2/virt/2.2/bin/mpispawn 0
>
>
> I can reach both hosts from each other and they see the same IP
> addresses, and I can ssh to each of the two hosts from each of the
> two hosts.
>
>
> Could you help me to identify the issue?
>
> The MVAPICH-Virt version is the last accessible on the mvapich
> download page
>
> P. S.
>
> You say to allow reading for the user in section 5 of the Virt user
> manual, but actually the library asks for write access as well, so I
> set up rights to 766.
>
> --
> Regards,
> Maksym Planeta
>
> _______________________________________________
> mvapich-discuss mailing list
> mvapich-discuss at cse.ohio-state.edu
> <mailto:mvapich-discuss at cse.ohio-state.edu>
> http://mailman.cse.ohio-state.edu/mailman/listinfo/mvapich-discuss
>
> --
> Thanks,
> -Jie
>
> --------------------------------------------
> Department of Computer Science & Engineering
> Erwin E. Dreese Laboratories, 2015 Neil Avenue
> Ohio State University
--
Regards,
Maksym Planeta
-------------- next part --------------
A non-text attachment was scrubbed...
Name: smime.p7s
Type: application/pkcs7-signature
Size: 5174 bytes
Desc: S/MIME Cryptographic Signature
URL: <http://mailman.cse.ohio-state.edu/pipermail/mvapich-discuss/attachments/20170705/e00a15bc/attachment.p7s>
More information about the mvapich-discuss
mailing list