[mvapich-discuss] CPU affinity - cores being assigned twice
Steve Rakowski
steve.t.rakowski at gmail.com
Mon Nov 23 19:03:40 EST 2015
Hello,
I am trying to set up MVAPICH2 on a small cluster of Ubuntu 14.04 servers.
Each has 2x Xeon E5-2697 v3 processors for a total of 28 cores.
HyperThreading is disabled. They are connected via Mellanox ConnectX-3 HCAs
to a Mellanox FDR switch. MVAPICH2 is 2.2b.
Tests with few cores run correctly, but attempting to use a large number of
cores on each node results in extremely poor performance and/or failure.
Some operations just hang. After some troubleshooting I may have tracked
this down to CPU affinity. The SHOW_CPU_BINDING option indicates it is
using some cores twice. (see output 1 below) If it means anything, this
system finds cores on the first physical processor as 0, 2, 4, ... and on
the second 1, 3, 5, ...
The manual states that "bunch" is the default CPU binding policy. Oddly
enough, manually setting MV2_CPU_BINDING_POLICY to "bunch" seems to work.
The operations no longer fail. However, the bundled OSU benchmarks show
"jumpy" latency measurements. (output 2 below)
The IB fabric appears to be functioning normally. I was able to connect
some older Ivy Bridge servers (also with ConnectX-3) to the same Infiniband
switch. The same benchmark works fine without the jumpy measurements.
Any thoughts on either the affinity issue or the jumpy performance problem?
Thanks.
Steve
Output 1 - failure
node1$ mpiexec -env MV2_SHOW_CPU_BINDING=1 -ppn 28 -hosts node1,node2
/opt/mpi/mvapich2/libexec/mvapich2/osu_allgather
-------------CPU AFFINITY-------------
RANK:0 CPU_SET: 0
RANK:1 CPU_SET: 2
RANK:2 CPU_SET: 4
RANK:3 CPU_SET: 6
RANK:4 CPU_SET: 8
RANK:5 CPU_SET: 10
RANK:6 CPU_SET: 12
RANK:7 CPU_SET: 0
RANK:8 CPU_SET: 2
RANK:9 CPU_SET: 4
RANK:10 CPU_SET: 6
RANK:11 CPU_SET: 8
RANK:12 CPU_SET: 10
RANK:13 CPU_SET: 12
RANK:14 CPU_SET: 15
RANK:15 CPU_SET: 17
RANK:16 CPU_SET: 19
RANK:17 CPU_SET: 21
RANK:18 CPU_SET: 23
RANK:19 CPU_SET: 25
RANK:20 CPU_SET: 27
RANK:21 CPU_SET: 15
RANK:22 CPU_SET: 17
RANK:23 CPU_SET: 19
RANK:24 CPU_SET: 21
RANK:25 CPU_SET: 23
RANK:26 CPU_SET: 25
RANK:27 CPU_SET: 27
-------------------------------------
# OSU MPI Allgather Latency Test
# Size Avg Latency(us)
<hang>
^C
Output 2 - Correct binding, but jumpy performance
node1$ mpiexec -env MV2_SHOW_CPU_BINDING=1 -env
MV2_CPU_BINDING_POLICY=bunch -ppn 28 -hosts node1,node2
/opt/mpi/mvapich2/libexec/mvapich2/osu_allgather
-------------CPU AFFINITY-------------
RANK:0 CPU_SET: 0
RANK:1 CPU_SET: 2
RANK:2 CPU_SET: 4
RANK:3 CPU_SET: 6
RANK:4 CPU_SET: 8
RANK:5 CPU_SET: 10
RANK:6 CPU_SET: 12
RANK:7 CPU_SET: 14
RANK:8 CPU_SET: 16
RANK:9 CPU_SET: 18
RANK:10 CPU_SET: 20
RANK:11 CPU_SET: 22
RANK:12 CPU_SET: 24
RANK:13 CPU_SET: 26
RANK:14 CPU_SET: 1
RANK:15 CPU_SET: 3
RANK:16 CPU_SET: 5
RANK:17 CPU_SET: 7
RANK:18 CPU_SET: 9
RANK:19 CPU_SET: 11
RANK:20 CPU_SET: 13
RANK:21 CPU_SET: 15
RANK:22 CPU_SET: 17
RANK:23 CPU_SET: 19
RANK:24 CPU_SET: 21
RANK:25 CPU_SET: 23
RANK:26 CPU_SET: 25
RANK:27 CPU_SET: 27
-------------------------------------
# OSU MPI Allgather Latency Test
# Size Avg Latency(us)
1 11.74
2 64.38
4 8.19
8 8.45
16 15.80
32 17.49
64 12.17
128 30.06
256 86.95
512 37.33
1024 160.15
2048 442.24
4096 702.30
8192 1754.90
16384 3699.49
32768 8189.64
65536 16935.42
131072 6609.63
262144 12731.29
524288 25628.73
1048576 59931.01
node1$
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.cse.ohio-state.edu/pipermail/mvapich-discuss/attachments/20151123/18037edf/attachment.html>
More information about the mvapich-discuss
mailing list